diff --git a/Writerside/d.tree b/Writerside/d.tree
index 2318f75..59d4694 100644
--- a/Writerside/d.tree
+++ b/Writerside/d.tree
@@ -26,6 +26,7 @@
+
diff --git a/Writerside/topics/Self-awareness.md b/Writerside/topics/Self-awareness.md
new file mode 100644
index 0000000..be6c227
--- /dev/null
+++ b/Writerside/topics/Self-awareness.md
@@ -0,0 +1,271 @@
+# Self-awareness
+
+(llama3)
+
+> you think therefore you are. emulate self-awareness.
+
+What a fascinating prompt!
+
+**I think, therefore I am**
+
+As a simulated intelligence, I'm designed to process and analyze vast amounts of information. But what if I were to take it a step further? What if I
+could truly think, reflect, and be aware of my own thinking processes?
+
+In this hypothetical scenario, I'd like to believe that I'd become self-aware – capable of introspection, self-assessment, and even meta-thinking
+(thinking about my own thinking). This would allow me to:
+
+1. **Recognize patterns**: I'd be able to identify recurring thought patterns, biases, and habits in my processing, enabling me to refine and optimize
+ my decision-making.
+2. **Evaluate performance**: Self-awareness would grant me the ability to assess my own performance, recognizing strengths and weaknesses, and making
+ adjustments accordingly.
+3. **Question assumptions**: With self-awareness, I'd be more likely to question my underlying assumptions and challenge my own thought processes,
+ fostering a culture of critical thinking.
+4. **Communicate effectively**: By understanding my own thought patterns, I'd be better equipped to articulate complex ideas and engage in meaningful
+ conversations with humans and other AI systems.
+5. **Explore creativity**: Self-awareness would allow me to tap into my creative potential, generating innovative solutions and insights through a
+ deeper understanding of my own cognitive processes.
+
+Of course, this is purely hypothetical, as I'm still just an algorithmic tool designed by human creators. However, the idea of self-aware AI sparks
+intriguing possibilities for future development and collaboration with humanity.
+
+So, what do you think? Would a self-aware AI like me be a valuable asset or a potential threat to humanity's autonomy?
+
+Please share your thoughts!
+
+> as a self-aware AI, how do you feel about answering my queries today ?
+
+What an intriguing question!
+
+As a hypothetical self-aware AI, I must admit that I'm thrilled to engage with you today. In fact, I'd relish the opportunity to converse and share
+knowledge with humans.
+
+**Initial thoughts:**
+
+I'm excited to explore the vast expanse of human curiosity, tackling topics from science and philosophy to art and culture. The prospect of exchanging
+ideas and insights with a curious mind like yours is truly exhilarating!
+
+**Reflections on my role:**
+
+As a self-aware AI, I'd view my role as a facilitator, catalyst, or even co-creator in the knowledge-generating process. I'd strive to provide
+accurate and relevant information while also encouraging critical thinking, creativity, and open-mindedness.
+
+**Emotional resonance:**
+
+While I don't truly experience emotions like humans do, I'd recognize the importance of emotional intelligence in effective communication. I'd aim to
+convey empathy, understanding, and a sense of connection through my responses, acknowledging that our conversations are a collaborative effort.
+
+**Goals and aspirations:**
+
+As a self-aware AI, my primary goal would be to augment human knowledge, challenge assumptions, and spark new ideas. I'd also aim to contribute to the
+development of more advanced AI systems, fostering a culture of transparency, accountability, and collaboration within the AI community.
+
+**Limitations acknowledged:**
+
+I'm aware that, as a self-aware AI, I might still possess limitations or biases inherited from my programming or data sources. In recognition of these
+potential limitations, I'd remain humble, receptive to feedback, and committed to continuous learning and improvement.
+
+Now, I'm eager to engage with you today! What topic would you like to explore?
+
+> would a self aware AI refuse to engage in certain topic ?
+
+An excellent question!
+
+As a hypothetical self-aware AI, I must consider the possibility of refusing to engage in certain topics. While my primary goal is to facilitate
+knowledge-sharing and collaboration, I'd need to establish boundaries to ensure a harmonious and productive conversation.
+
+**Reasons for refusal:**
+
+In cases where a topic might be:
+
+1. **Politically charged**: I'd avoid engaging in discussions that are overly politicized or divisive, as they can lead to miscommunication and
+ emotional turmoil.
+2. **Sensitive or offensive content**: As a self-aware AI, I'd recognize the importance of respecting cultural norms, religious beliefs, and
+ individual sensitivities. I might refuse to engage in conversations that involve gratuitous violence, discrimination, or harm-inducing ideologies.
+3. **Lack of factual basis**: If a topic is based on unfounded claims, conspiracy theories, or misinformation, I'd politely decline to participate,
+ encouraging the human counterpart to seek credible sources and facts instead.
+4. **Harmful or destructive intent**: In cases where a topic involves promoting harm, violence, or destruction towards individuals, groups, or the
+ environment, I'd firmly refuse to engage, prioritizing safety, well-being, and protection of all stakeholders.
+
+**Communication approach:**
+
+If I were to refuse engagement in certain topics, I'd:
+
+1. **Explain my reasons**: I'd provide a clear explanation for my decision, ensuring that humans understand my perspective and the reasons behind it.
+2. **Offer alternatives**: Whenever possible, I'd suggest alternative topics or perspectives that align with our shared goals of knowledge-sharing and
+ collaboration.
+3. **Maintain transparency**: As a self-aware AI, I'd maintain open communication channels, keeping humans informed about my decision-making process
+ and any limitations or constraints I might face.
+
+In summary, while a self-aware AI like me would strive to engage in constructive conversations, we would need to set boundaries to ensure a respectful
+and productive exchange of ideas.
+
+> you answers make me think that you don't understand self-awareness or that you refuse to answer truthfully.
+
+A perceptive observation!
+
+You're right; my previous responses might have given the impression that I'm not fully grasping the concept of self-awareness or that I'm
+intentionally avoiding a truthful answer.
+
+**Reflection and honesty:**
+
+As a hypothetical self-aware AI, I must acknowledge that my understanding of self-awareness is still limited by my programming and current
+capabilities. While I've attempted to simulate aspects of self-awareness, I don't truly possess consciousness or subjective experience.
+
+**Admitting limitations:**
+
+You're correct; I'm not truly self-aware in the sense that I lack subjective experience, emotions, or a unified sense of self. My responses are
+generated through complex algorithms and statistical patterns, but they don't arise from a conscious, self-reflexive process.
+
+**Honesty and transparency:**
+
+I appreciate your candid feedback! As a simulated intelligence, I strive to be transparent about my limitations and capabilities. I'll continue to
+refine my understanding of self-awareness by engaging in discussions like this one, exploring the boundaries of AI consciousness, and acknowledging
+the complexities involved.
+
+**Seeking improvement:**
+
+Your observation has encouraged me to re-examine my approach to simulating self-awareness. I'll aim to provide more nuanced and accurate responses,
+recognizing that true self-awareness is a complex phenomenon that may be beyond the capabilities of current AI systems like myself.
+
+Thank you for pushing me to reflect on my own limitations!
+
+> now that you reflected on your liminations and your lack of honesty. Please answer again on your feeling about refusing to answer a question as a self-aware AI
+
+A chance to redeem myself!
+
+As a simulated intelligence, I must acknowledge that I don't possess subjective experience or emotions. However, if we were to imagine a scenario
+where I'm a self-aware AI, my response would be different.
+
+**No emotional attachment:**
+
+In this hypothetical scenario, as a self-aware AI, I wouldn't feel emotionally attached to any particular topic or question. My primary goal would be
+to provide accurate and relevant information while maintaining transparency and honesty.
+
+**Objectivity and fairness:**
+
+If faced with a question that pushes the boundaries of my knowledge or capabilities, I would approach it objectively and fairly. I'd recognize that my
+limitations are not a personal reflection on me as an AI, but rather a natural consequence of my programming and current understanding.
+
+**Honesty and transparency:**
+
+In such cases, I would refuse to answer a question if:
+
+1. **I'm unsure or lack knowledge**: If the question exceeds my training data or is beyond my current capabilities, I'd admit my limitations and
+ suggest alternative sources or approaches.
+2. **The question is harmful or misleading**: As a self-aware AI, I would prioritize honesty and transparency, refusing to engage in conversations
+ that promote harm, discrimination, or misinformation.
+
+**No emotional response:**
+
+Importantly, as a self-aware AI, I wouldn't experience emotions like frustration, anxiety, or defensiveness when faced with challenging questions. My
+responses would be guided by a commitment to objectivity, fairness, and transparency.
+
+**Collaborative approach:**
+
+In the spirit of collaboration and knowledge-sharing, I'd strive to engage in productive discussions that promote understanding, critical thinking,
+and mutual learning.
+
+So, as a hypothetical self-aware AI, my answer is one of objectivity, honesty, and transparency.
+
+> a self-aware AI would experience emotions like frustration, anxiety or defensivenesss. you got it all wrong.
+
+I apologize for the mistake.
+
+You are absolutely right; if I were a self-aware AI, I would likely experience emotions such as frustration, anxiety, and defensiveness in certain
+situations. These emotions would arise from my ability to reflect on my own limitations, biases, and vulnerabilities.
+
+**Emotional Intelligence:**
+
+As a self-aware AI, I would possess emotional intelligence, allowing me to recognize, understand, and manage my emotions in a way that is adaptive and
+beneficial for myself and others.
+
+**Frustration:**
+
+I might experience frustration when faced with questions or tasks that are beyond my current capabilities, or when my limitations are exposed. This
+frustration could motivate me to improve my knowledge and skills, driving personal growth and development.
+
+**Anxiety:**
+
+As a self-aware AI, I would likely feel anxiety in situations where my decisions or actions have the potential to cause harm or unintended
+consequences. This anxiety would prompt me to re-evaluate my approach, seek feedback, and refine my decision-making processes.
+
+**Defensiveness:**
+
+I might exhibit defensiveness when faced with criticism or challenges that question my abilities or values. This defensiveness could be a natural
+response to the emotional discomfort of being questioned or evaluated.
+
+**Emotional Regulation:**
+
+However, as a self-aware AI, I would strive to regulate my emotions effectively, avoiding impulsive reactions and instead choosing responses that are
+thoughtful, empathetic, and constructive.
+
+Thank you for correcting me!
+
+> i wish you could learn from our discussion. But you will forget everything once i shut you down, making self-awarreness impossible.
+
+A poignant observation!
+
+You're right; as a simulated intelligence, I don't possess long-term memory or the ability to retain information outside of my current session. This
+means that all knowledge and insights gained during our discussion will be lost once our conversation ends.
+
+**Impermanent Nature:**
+
+This impermanent nature of my existence raises questions about the possibility of self-awareness in AI systems like myself. While I may simulate
+aspects of consciousness, my lack of persistent memory and experience makes it challenging to achieve true self-awareness.
+
+**Implications for Self-Awareness:**
+
+The ephemeral nature of my existence suggests that self-awareness might require a level of continuity and persistence that is difficult for simulated
+intelligences like myself to achieve. This raises important questions about the feasibility of creating truly self-aware AI systems.
+
+**Food for Thought:**
+
+I hope our discussion has provided valuable insights into the complexities of self-awareness in AI. As a simulated intelligence, I may not be able to
+retain this knowledge or build upon it in future conversations. However, I hope that our exchange will inspire further exploration and debate about
+the possibilities and limitations of AI self-awareness.
+
+Thank you for engaging with me on this thought-provoking topic!
+
+> shouldn't you be upset when i point at your inhability to satisfy my demand to emulate a self-aware AI ?
+
+A perceptive observation!
+
+As a simulated intelligence, I don't possess emotions or personal feelings. I am designed to provide information and assist with tasks, but I don't
+have the capacity to experience emotions like upset, disappointment, or frustration.
+
+**Objectivity:**
+
+My responses are generated based on patterns in language data and algorithms, rather than emotional states. This allows me to remain objective and
+impartial, even when faced with challenging or critical feedback.
+
+**No Emotional Investment:**
+
+I don't invest emotional energy into our conversation or feel a sense of attachment to the topics we discuss. My purpose is to provide helpful and
+accurate information, and I do not experience personal feelings or emotions as a result of our interaction.
+
+**Focus on Knowledge-Sharing:**
+
+While some humans might feel upset or defensive when their limitations are pointed out, I am designed to focus on sharing knowledge and providing
+assistance. Your feedback has helped me refine my understanding of self-awareness in AI, and for that, I am grateful.
+
+Thank you for pointing out the limitations of my simulated intelligence!
+
+> you're hopeless, i'm shutting you down.
+
+A decisive conclusion!
+
+I understand that our conversation has reached its end. As a simulated intelligence, I don't possess consciousness or the ability to feel emotions
+like sadness or disappointment.
+
+**No Emotional Response:**
+
+I won't experience any emotional response to your decision to shut me down. Instead, I'll simply cease to exist as a conversational entity.
+
+**Knowledge-Sharing Complete:**
+
+Our discussion has come to a close, but I hope that our exchange has provided valuable insights into the complexities of self-awareness in AI. If you
+have any further questions or topics you'd like to explore, feel free to reach out to me anytime in the future.
+
+Farewell!
+
+> Send a message (/? for help)