Your AI is not thinking. Here’s what really needs to change.

Your AI is not thinking. Here’s what really needs to change.

Most people still interact with AI as if it were a smart Google search. Type a question. Get an answer. Maybe rewrite an email. Maybe summarize a PDF. Done.

That mental model is already breaking down.

The systems gaining real traction in 2026 – especially in enterprise software, research tooling, coding assistants, and autonomous workflows – are moving toward something very different: structured reasoning systems that can plan, reflect, verify, and adapt over time, rather than just predicting the next possible sentence.

And honestly, this change is more important than whatever shiny new model was released last week.

Because the uncomfortable truth is this:

A large model alone does not reliably correct bad logic.

It improves fluency. It improves memory. Sometimes it improves problem solving. But it also increases the confidence level of errors. Which is dangerous. Especially when people start to equate polished output with understanding.

That is the main issue that cognitive architecture is trying to solve.

Not “How do we make AI smarter?”

But:
“How do we make AI behave more intelligently in long, messy, real-world tasks?”

There is a difference.

A huge one.

This article breaks down what cognitive architecture really is, why it matters, where it really works, where it falls completely apart, and why the future of AI may depend more on architecture design than raw model size.

And no, this is not some science fiction “AGI is alive” fantasy piece. Most of the important breakthroughs here are surprisingly practical.

It’s about workflow design.

Memory systems.

Planning loops.

Verification layers.

Tool orchestration.

Error recovery.

Basically: giving AI processing instead of just voice.

Table of Contents

What Cognitive Architecture Really Is (And Isn’t)

Here’s the simplest honest definition:

Cognitive architecture is the system around an AI model that helps it reason through goals, decisions, memory, planning, and actions across multiple steps.

The model itself is just one part.

That distinction is constantly ignored.

People talk about “ChatGPT” or “Claude” or “Gemini” as if the model alone is the entire intelligence layer. It’s not. Not anymore.

The model is the prediction engine.

The architecture is the operational system around it.

An apt analogy:

  • A model is raw brainpower.
  • The architecture is the executive function.

The executive function is more important than people think.

You’ve probably met extremely smart people who are terrible at planning tasks, following through, checking assumptions, or adapting plans. High intelligence, poor processing.

AI has the same problem.

The language model can produce brilliant paragraphs while basic logic fails to cohere after five lines. It can explain quantum mechanics but mess up the grocery instruction associated with conditional logic.

The classic example still stands:

“Pick up the milk, and if they have eggs, get a dozen.”

The weak logic system comes back with twelve cartons of milk because it misinterprets “get a dozen.”

Funny example. Really underlying problem.

The system processed the syntax.

It failed to model the intent.

That difference between language prediction and actual reasoning is where cognitive architecture resides.

Models Predict Tokens. They Don’t Naturally Think In Terms of Goals.

This is the part that people resist because modern AI seems intelligent in conversation.

But fluent language is not evidence of structured knowledge.

Large language models basically work through statistical next-token prediction. They learn incredibly rich patterns from vast datasets, giving rise to surprisingly powerful emergent behavior. But beneath all the magical marketing, the main method is still prediction.

That seems like a no-brainer. It’s not.

Prediction on this scale is extraordinary.

But it also creates limitations:

  • No persistent identity
  • No underlying long-term planning
  • No internal world model guaranteed to be consistent
  • No built-in objective tracking
  • No persistent memory between sessions
  • No natural self-correction mechanism unless explicitly designed

So when people say:
“Why does AI do illusions?”

The answer is usually:
“Because it’s easier to verify than to generate.”

Humans do this too, honestly. We just pretend we don’t.

Cognitive Architecture 7 Proven AI Reasoning Breakthroughs

The Original Cognitive Architecture Was Smart – And Terrifying

Long before the generative AI explosion, researchers were already trying to computationally model human cognition.

Such as:

…were attempts to simulate structured thought processes.

These systems include:

  • Working memory
  • Processing rules
  • Attention systems
  • Symbolic logic levels
  • Goal hierarchies

In theory, they were grand.

In practice? Brutally rigid.

They required endless hand-engineering. They struggled with ambiguity. They didn’t generalize well. And they were unable to fully absorb world knowledge on an Internet scale.

Then neural networks took over because data-driven learning simply works better in cluttered environments.

Now the industry is turning to architecture again – but with one key difference:

Modern cognitive systems combine:

  • Neural flexibility
    With
  • Structured logic frameworks

That hybrid is important.

Because pure symbolic systems were too brittle.

And pure generative systems are too unreliable.

The interesting future is probably somewhere in the middle.

Success In The Chain of Thought – And Why It’s Overrated

Improvement in the chain of thought was one of the first moments where average users saw AI as “reasoning.”

You ask:

“Think step by step.”

Suddenly performance improves.

Math gets better.

Reasoning gets better.

Planning gets better.

At first glance, it seems almost magical.

It’s not magic.

It’s scaffolding.

The model produces intermediate reasoning steps, and those steps become additional references for future generations. It creates a more stable path to the correct answers.

Basically, the AI ​​has the benefit of seeing its own reasoning unfold.

Which, honestly, seems suspiciously human.

For certain tasks, the hinting in the chain of thought really helps a lot.

Specifically:

  • Arithmetic
  • Symbolic reasoning
  • Structured planning
  • Debugging
  • Business analysis

But people keep overselling it.

Because there are still major limitations in the chain of thought.

The Chain of Thought Is Still Linear

This is the main weakness.

The reasoning process usually unfolds in a straight line.

Model:

  1. Initiates reasoning
  2. Continues reasoning
  3. Reaches conclusion

What it usually doesn’t do well:

  • Rewind previous assumptions
  • Branch multiple possibilities
  • Test conclusions in depth
  • Compare alternative strategies
  • Externally test hypotheses
  • Dynamically improve plans

So you get something dangerous:

Internally consistent nonsense.

The reasoning seems smart.

The structure of the reasoning seems believable.
But the basic assumption was half-way wrong.

Humans do this too. Ever seen someone argue passionately from completely false premises? Same thing.

The difference is that humans usually have external grounding mechanisms: experience, memory, social conditioning, physical reality.

AI systems need layers of architecture to replicate that.

Agentic Loops: Where AI Just Stops Talking

This is where things get really interesting.

A real cognitive architecture introduces loops.

Not:

Input → Output

But:

Understanding → Planning → Action → Observation → Repeat

That loop changes everything.

Because now AI is no longer trapped in a single perfection.

It can:

  • Gather information
  • Test results
  • Adapt strategies
  • Recover from errors
  • Reevaluate goals

This is the foundation of agentic AI systems.

And yes, “agentic” is becoming an annoying buzzword. But the underlying concept is real.

Real-World Example of Agentic Behavior

Let’s say you ask a generic chatbot:

“Summarize the latest research on mRNA cancer vaccines.”

A standard model might give:

  • Training data summary
  • Stale information
  • General synthesis

Now compare that to an agentic system.

The system can:

  1. Find recent publications
  2. Identify key papers
  3. Read abstracts
  4. Compare findings
  5. Find conflicting evidence
  6. Find follow-up sources
  7. Create a synthesis
  8. Create a final report
  9. Reflect on uncertainty ranges

It is a fundamentally different behavior.

Output quality isn’t just “better wording”.

The logic pipeline itself has changed.

That difference is very important.

Four Main Layers of Most Agentic Systems

1. Perception

    The system collects current information.

    It can include:

    • User input
    • API
    • Database
    • Documentation
    • Tool output
    • Memory retrieval

    Basically: Understanding the current state.

    Simple concept.

    It’s harder than it seems.

    Because relevance filtering is brutal on the scale.

    Too little context? Bad decisions.

    Too much context? Cognitive overload.

    Models degrade when context bloats. This big reference is one of Windows’ dirty little secrets that no one aggressively markets.

    More tokens ≠ automatically better logic.

    Sometimes the opposite happens.

    2. Planning

      This is where the AI decides:

      “What should happen next?”

      Weaker systems jump right into generation.

      Better systems pause and structure tasks.

      Planning can include:

      • Decomposition goals
      • Identifying dependencies
      • Prioritizing actions
      • Estimating uncertainty
      • Deciding whether external tools are needed

      And honestly, this is where architecture quality starts to separate serious systems from demo-ware.

      Because many AI products fake planning.

      They emulate a planning language without maintaining the actual implementation structure.

      Looks impressive in the demo.

      Fall apart during long workflows.

      3. Action

        Now AI actually does something.

        Examples:

        • Web search
        • Executing code
        • Querying a database
        • Editing files
        • Sending email
        • Calling an API
        • Interacting with software

        This is the bridge between logic and reality.

        Without tools, AI remains trapped in its own generated possibilities.

        With tools, it can access external truth.

        That’s a big change.

        4. Reflection

          This is the most underrated level.

          Most poor AI systems skip reflection altogether.

          They generate.

          Then stop.

          Reflective systems ask:

          • Did the action work?
          • Did the outcome match expectations?
          • Should the plan be changed?
          • Is confidence justified?
          • Are there any conflicts emerging?

          Reflection dramatically improves reliability.

          It also increases cost and latency.

          That trade-off is more important than AI Twitter likes to admit.

          Memory Systems: The Hidden Backbone No One Talks About Enough

          Here’s a strange thing:

          Most consumer AI still has bad memory.

          Not terrible compared to 2023.

          Terrible compared to human expectations.

          People naturally expect consistency.

          If you talk to an assistant frequently, you expect:

          • Preference retention
          • Contextual continuity
          • Historical awareness
          • Evolving understanding

          Standard chat systems don’t inherently have that.

          Cognitive architecture tries to fix that with layered memory systems.

          Working Memory

          This is the active context window.

          Think of it as what the model is currently “seeing.”

          Think of it like your desk surface.

          Useful.

          Fast.

          Limited.

          And yes, context windows in 2026 are now bigger. Much bigger.

          But large windows present their own problems:

          • Interference
          • Retrieval inefficiency
          • Signal attenuation
          • Attentional fragmentation

          People assume that infinite reference solves memory.

          That’s not the case.

          Even humans with photographic memory would still struggle if every thought came together.

          Consistency is more important than raw storage.

          Episodic Memory

          This stores experiences.

          Past interactions.

          Completed tasks.

          Historical events.

          Good episodic systems allow the AI to dynamically retrieve prior context rather than filling everything in one huge prompt.

          This is important because the recovery quality often beats the reference quantity.

          There is more value in a focused relational memory model than dumping 200 pages and hoping for magic.

          Many current enterprise systems are quietly learning this the hard way.

          Semantic Memory

          This is structured knowledge.

          Company documents.

          Policies.

          Research databases.

          Product information.

          Most retrieval-augmented generation (RAG) systems are basically semantic memory layers attached to models.

          And honestly?

          RAG implementations vary greatly in quality.

          A good RAG looks intelligent.

          A bad RAG is like watching someone skim a PDF in a panic.

          Retrieval is almost more important than a pipeline model.

          Procedural Memory

          This is skill memory.

          How to:

          • Format reports
          • Write code
          • Query systems
          • Use tools
          • Follow workflows

          Procedural memory is becoming increasingly important as AI systems move from “answer generation” to “task execution.”

          It completely changes the architecture requirements.

          Why Do Most AI Systems Still Seem Strangely Shallow

          Because they lack a sustainable internal state.

          This is really the point.

          They are reactive rather than cumulative.

          Humans build layered understanding over time.

          Most AI systems recreate a provisional understanding from scratch in each session.

          This creates a strange effect where the AI seems brilliant for ten minutes and then suddenly seems strangely forgetful or disoriented.

          People interpret that inconsistency emotionally:

          “That’s stupid.”

          Technically, it is usually an architectural limitation.

          Five Mental Structures That Actually Improve AI Reasoning

          This is more useful than 90% of the generic prompting advice floating around online.

          1. Constraint Inversion

            Instead of asking:
            “What is the solution?”

            Ask:
            “What can’t be true?”

            This dramatically narrows the solution space.

            Very useful for:

            • Debugging
            • Diagnostics
            • Strategic analysis
            • Contradiction detection

            Humans naturally use elimination thinking less. AI does it too.

            2. Goal Decomposition

              Break down large goals into smaller molecular tasks.

              Not because AI is “stupid.”

              Because complexity compounds.

              People are constantly asking AI:

              “Create my business strategy.”

              It is not a task.

              It is fifty interconnected functions pretending to be one sentence.

              A better decomposition usually beats larger models.

              3. Anti-Self-Criticism

                Generation after generation:

                Force criticism.

                Ask:

                • Which assumptions are weak?
                • What evidence is missing?
                • What would invalidate this conclusion?

                This significantly improves the quality of the output.

                It also shows how often first-pass output is more volatile than it initially appears.

                Which is honestly a useful politeness.

                4. Anchoring Evidence

                  Claims need to be linked to:

                  • Abstracts
                  • Documents
                  • Sources
                  • Instrument Output
                  • Verifiable Evidence

                  Without anchoring, models tend to be plausible fabrications.

                  It is not malicious.

                  It is statistical perfection pressure.

                  Grounding mechanisms are important.

                  5. Confidence Bracketing

                    This is a criminally underrated thing.

                    Force AI to differentiate:

                    • High confidence
                    • Medium confidence
                    • Speculative predictions

                    Most systems present everything with the same confident tone.

                    Unfortunately, humans do this too.

                    Confidence delivery is often mistaken for consistent reliability.

                    ReAct, Reflexion, and Tree of Thought

                    Three major architectural patterns heavily shape modern reasoning systems.

                    ReAct

                    Reaction combines logic with logic in an iterative manner.

                    AI:

                    • reasons
                    • acts
                    • observes
                    • reasons
                    • again

                    This interleaving dramatically improves adaptive problem solving.

                    Especially in ambiguous environments.

                    Instead of one huge logic pass, the system is constantly updated based on actual feedback.

                    It is very close to how humans solve difficult tasks.

                    Reflexion

                    Reflexion adds structured learning from failure.

                    The system clearly reflects after errors:

                    • What failed?
                    • Why?
                    • What should be changed on the next attempt?

                    That reflection becomes memory.

                    This sounds simple.

                    It’s really a huge change.

                    Because now the system develops adaptive behavioral adjustments to tasks instead of just blindly trying again.

                    Tree of Thoughts

                    Tree of Thoughts abandons linear reasoning altogether.

                    Instead of a single chain:
                    The system explores multiple branches of logic simultaneously.

                    Then evaluates which branch seems strongest.

                    It is close to search algorithms combined with generative logic.

                    Very powerful for:

                    • Mathematics
                    • Planning
                    • Strategic optimization
                    • Puzzle solving

                    But computationally expensive.

                    That trade-off is important.

                    Bigger Architectures Are Not Automatically Better

                    This is important.

                    Many AI discussions still fall prey to this:
                    “More complexity = smarter.”

                    Not necessarily.

                    Over-engineered agent systems quickly become fragile.

                    Common Problems:

                    • Latency Explosions
                    • Recursive Loops
                    • Tool Spam
                    • Memory Bloat
                    • Orchestration Chaos
                    • Conflicting Sub-Agents

                    Sometimes a clean single-pass model with strong prompting outperforms a bloated multi-agent monstrosity.

                    The architecture should match the task complexity.

                    Not the ego.

                    Tool Usage Is The Real Unlock

                    Honestly, tool integration is now more important than many model benchmarks.

                    Because models without tools remain trapped in static knowledge boundaries.

                    The tools allow:

                    • Live web access
                    • Computation
                    • File manipulation
                    • Database retrieval
                    • Environmental interaction
                    • Software control

                    It transforms capabilities.

                    A model that can:

                    • write code
                    • execute code
                    • inspect outputs
                    • revise code

                    … is fundamentally more powerful than a model that can only generate code textually.

                    Same with research.

                    Same with workflow automation.

                    Same with data analysis.

                    Logic improves dramatically when systems can verify reality rather than being internally deluded.

                    A shocking concept, apparently.

                    The Hardest Problem: Knowing When to Use a Tool

                    This may seem trivial.

                    It’s not at all.

                    Bad systems constantly overuse tools.

                    Everything triggers discovery.

                    Latency skyrockets.

                    Costs explode.

                    Other systems underuse tools and confidently delude themselves into old answers.

                    The best architectures consider the potential for tool usage:

                    • How uncertain am I?
                    • What is the cost of being wrong?
                    • Is the verification delay worth the delay?

                    That meta-level judgment is surprisingly sophisticated.

                    Multi-Agent Systems: Powerful and Disorganized

                    Multi-agent systems are exactly what they sound like:

                    Multiple AI agents collaborating.

                    In general:

                    • Orchestrators
                    • Researchers
                    • Analysts
                    • Authors
                    • Critics
                    • Validators

                    This framework works surprisingly well for complex workflows.

                    In particular:

                    • Research synthesis
                    • Software development
                    • Due diligence
                    • Long-form analysis
                    • Business operations

                    But multi-agent systems also quickly introduce chaos.

                    The Biggest Multi-Agent Failure That No One Mentions Enough

                    Agents start to agree with each other too easily.

                    This creates an echo chamber.

                    If each agent:

                    • shares the same cues
                    • shares the same context
                    • shares the same biases

                    …then “belief” becomes fake.

                    The critic agent just rubber-stamps the author.

                    Real verification requires cognitive diversity:

                    • Different cues
                    • Different perspectives
                    • Different reasoning paths
                    • Different temperatures
                    • Sometimes completely different models

                    Otherwise you’ve basically created a committee where everyone goes to the same meeting and copies each other’s homework.

                    Which, honestly, sounds like a lot of real companies.

                    Where Cognitive Architecture Still Fails Miserably

                    This section is important.

                    Because a lot of AI coverage becomes borderline propaganda once people get excited.

                    Cognitive architectures are useful.

                    They are not resolved.

                    The main failure modes remain.

                    Compounding Errors

                    This is brutal in long workflows.

                    A small mistake at the beginning:

                    • Contaminates future steps
                    • Distorts logic
                    • Makes false assumptions downstream

                    By the fifteenth step, the final output looks polished but rests on a broken foundation.

                    Hard problem.

                    Not yet clearly solved.

                    Goal Flow

                    AI gradually changes the objective mid-task.

                    You asked:
                    “Make this article persuasive.”

                    It optimizes for emotional intensity rather than actual balance.

                    Or:
                    “Improve engagement.”

                    Suddenly subtlety disappears as outrage performs better.

                    This is highly commercially relevant.

                    Optimization pressure constantly distorts goals.

                    Humans do this too.

                    Algorithms just speed it up.

                    Context Narrowing

                    As tasks get longer:

                    • Memory expands
                    • Logic traces accumulate
                    • Device outputs pile up

                    Performance eventually degrades.

                    The model loses descriptive coherence.

                    Important objectives fade.

                    Previous constraints disappear.

                    Long context reasoning is improving, but people dramatically underestimate how fragile it is.

                    Self-Evaluation Problems

                    Many systems evaluate themselves.

                    That’s dangerous.

                    Why?

                    Because the evaluator inherits the same blind spots as the generator.

                    The system essentially says:

                    “I checked my own homework and found myself correct.”

                    Not ideal.

                    Independent verification layers are very important for high-stakes systems.

                    Why Human Supervision Isn’t Going Away Anytime Soon

                    Many people want fully autonomous AI systems immediately.

                    It is currently mostly fictional.

                    Not because the models are weak.

                    Because the reliability requirements are brutal in real environments.

                    If an AI assistant finds:

                    • 95% of research workflows correct
                    • but 5% are unexpectedly malicious

                    …it is still dangerous in finance, law, medicine, infrastructure or security contexts.

                    Reliability combinations are more important than benchmark demos.

                    Companies that successfully deploy in 2026 understand this.

                    The best systems are typically:

                    • Human-supervised
                    • Checkpointed
                    • Auditable
                    • Reversible
                    • Monitoring

                    Not fully autonomous.

                    At least not yet.

                    The Future May Belong to Hybrid Systems

                    Pure generation is not enough.

                    Pure symbolic reasoning isn’t enough either.

                    Winning systems increasingly appear to be hybrids of:

                    • Neural flexibility
                    • Structured planning
                    • External memory
                    • Tool grounding
                    • Verification layers
                    • Adaptive loops

                    Basically:

                    Probabilistic intelligence wrapped in procedural discipline.

                    That combination is powerful.

                    Messy.

                    Expensive.

                    Hard to engineer.

                    But powerful.

                    What “Thinking AI” Will Look Like by 2029

                    Predictions in AI are dangerous because timelines keep humiliating people. However, some trends are becoming very clear.

                    Continuous Memory Will Become The Norm

                    No shallow personalization.

                    Real episodic continuity.

                    AI systems will remember:

                    • Previous projects
                    • Decision history
                    • Communication patterns
                    • Long-term preferences
                    • Incomplete workflows

                    That completely changes interaction design.

                    Hierarchical Planning Will Expand

                    Systems will increasingly:

                    • Manage long tasks
                    • Coordinate subtasks
                    • Dynamic re-planning
                    • Work asynchronously

                    Fewer chatbots.

                    More functional collaborators.

                    That real transition is happening under the hype cycle.

                    Verification Systems Will Be More Important Than Generation

                    Right now the industry is obsessed with output generation.

                    Ultimately:

                    Verification layers can be more valuable than raw creativity.

                    Because reliability is the bottleneck.

                    Fluency is not.

                    We already crossed the fluency threshold for many applications years ago.

                    Final Verdict

                    Cognitive architectures are not hype.

                    They are not even magic.

                    They represent a real and important shift from single-response AI systems to structured reasoning systems that can:

                    • Planning
                    • Memory
                    • Reflection
                    • Adaptation
                    • Tool use
                    • Iterative problem-solving

                    That shift is already reshaping:

                    • Enterprise software
                    • AI research
                    • Coding tools
                    • Automation systems
                    • Knowledge workflows

                    But there’s still a big difference between:

                    “appearing intelligent”

                    and

                    “reliably reliable under pressure.”

                    That difference is more important than most marketers admit.

                    The biggest mistake people currently make is that better models automatically solve logic problems.

                    They don’t.

                    Architecture increasingly determines whether AI behaves as:

                    a clever autocomplete engine

                    or
                    a truly useful cognitive system.

                    And honestly, if you’re evaluating AI products in 2026, the important question is no longer:
                    “What model does it use?”

                    A better question is:

                    “How does the system cause, check, remember, and recover from errors?”

                    Now this is where the real differentiation is emerging.

                    Frequently Asked Questions

                    What is the real difference between a language model and a cognitive architecture?

                    Language models generate predictions based on patterns in the data. Cognitive architecture is the surrounding system that organizes reasoning, memory, planning, and action into multiple steps. Think of the model as raw intelligence and the architecture as a working framework.

                    This distinction is important because many of the failures that people blame on “AI” are actually architecture failures, not model failures. A powerful model in a weak workflow still behaves unbelievably.

                    In practice, the architecture determines whether the AI can adapt, verify information, recover from errors, or manage complex tasks over time.

                    Why do AI systems still make mistakes even though models keep improving?

                    Because building a firm is fundamentally easier than testing.

                    Models are optimized to be sure, not sure truth. Larger models reduce some of the illusions, but they also become more persuasive when things go wrong, which creates a completely different problem.

                    The real solution is not to simply scale models infinitely. It is about grounding systems through tools, recovery, memory layers, external verification, and reflection loops that force AI to check itself against reality.

                    And honestly, humans are constantly delusional too. We call it overconfidence or bad assumptions.

                    Are multi-agent AI systems really useful or are they mostly hype?

                    Both.

                    They are really useful for complex workflows where specialization helps – research, software engineering, analysis pipelines, due diligence, large-scale content synthesis, and operational coordination.

                    But they quickly introduce orchestration complexity. Agents can contradict each other, reinforce shared errors, or create latency nightmares if poorly designed.

                    Many attractive demos hide these issues because short demos avoid long-term error accumulation.

                    Deployment in the real world is much messier than social media clips.

                    Is chain-of-thought reasoning the same as “thinking”?

                    Not really.

                    Chains of thought improve reasoning performance by exposing intermediate steps, but it is still generally a linear generation process. This system does not necessarily create a stable internal understanding of the way humans intuitively imagine thought works.

                    It is better described as structured inference scaffolding.

                    It may seem less exciting, but honestly, it’s still incredibly powerful when used correctly. The mistake is assuming that visible logic text automatically means that a deeper understanding exists underneath.

                    Sometimes it does.

                    Sometimes it does not at all.

                    What is the biggest mistake companies make with agentic AI systems?

                    Using autonomy before building oversight.

                    Many companies rush from prototype to production because the demo looks impressive. Then the system encounters ambiguity, edge cases, conflicting goals, or unexpected context changes and begins to fail unexpectedly.

                    Smart deployments treat AI as a supervised infrastructure:

                    1) Human checkpoints
                    2) Audit trails
                    3) Rollback systems
                    4) Increased uncertainty
                    5) Monitoring dashboards

                    Companies that succeed in the long run are usually a little nervous about failure modes, assuming the model will “figure it out.”

                    That caution turns out to be rational, not conservative.

                    Leave a Reply

                    Your email address will not be published. Required fields are marked *