DOI : https://doi.org/10.5281/zenodo.18802787
- Open Access

- Authors : Shaikh Abubakar Mosin, Memane Yadnyik Santosh, Sayyed Farid Navid, Ms. Sneha Sarode
- Paper ID : IJERTV15IS020537
- Volume & Issue : Volume 15, Issue 02 , February – 2026
- Published (First Online): 27-02-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
Integrating Multiple AI Models to Deliver Accurate & Reliable Responses
Shaikh Abubakar Mosin (23211830411), Memane Yadnyik Santosh (23213290154), Sayyed Farid Navid (23211830398)
Ms. Sneha Sarode
Department of Computer Engineering,Ā JSPMs Jayawantrao Sawant Polytechnic Hadapsar, Pune 411028
Abstract – In the fast-changing world of Artificial Intelligence (AI), large language models (LLMs) like Gemma, DeepSeek, Qwen, and Mistral have different strengths in areas like creative writing and technical analysis. However, answers to the same questions can differ widely in accuracy, depth, and clarity because of variations in training data, architectures, and goals. This paper introduces MetaMind, an AI-powered answer engine that combines several LLMs to provide accurate and reliable responses by comparing and ranking them. The system has a React-based frontend for easy user interaction and a Node.js backend for managing API requests, gathering responses, and ranking them. MetaMind evaluates answers based on accuracy, relevance, completeness, and clarity. It then chooses or combines the best response, reducing the drawbacks of relying on a single model. Testing highlights its usefulness in education, research, content creation, and general knowledge. This method not only improves the quality of responses but also demonstrates effective full-stack development and multi-AI management, setting the stage for future meta-AI systems.
Keywords – Large Language Models, Multi-LLM Integration, LLM Routing, Ensemble LLMs, Response Ranking, Query Answering, AI Orchestration
- INTRODUCTION
The rise of advanced AI models has changed how we retrieve information and resolve queries. Since the creation of the Transformer architecture in 2017 [Vaswani et al. 2017], large language models (LLMs) have grown significantly, with some exceeding 100 billion parameters by 2025. Models like Gemma (Google 2024), Qwen, and Mistral each excel in different areas. Gemma focuses on creative and interpretive tasks, DeepSeek handles technical and analytical reasoning, Qwen specialises in multilingual processing, and Mistral offers efficient, lightweight inference. This variety comes from differences in training datasets, architectural designs, such as transformer-based variants, and fine-tuning goals, leading to varied outputs for the same input queries. Despite these improvements, users still struggle to get consistent, high-quality responses. Depending on a single model may result in partial, biased, or incorrect answers, especially in crucial areas like education, research, and decision-making. Current platforms do not provide ways to compare and rank responses from multiple models, which creates a reliability gap. MetaMind solves this issue by building a system that queries several AI models at once, assesses their outputs using a structured ranking method, and delivers the best or combined response. This project uses complete technologies, including a responsive React frontend and a Node.js backend, to offer an easy-to-use, scalable solution. It was initially tested with models like ChatGPT, DeepSeek, and Blackbox, but has since evolved to include Gemma, DeepSeek, Qwen, and Mistral, showing its ability to adapt. This paper outlines the system’s architecture, methods, implementation, use cases, benefits, challenges, and future directions. It
emphasises contributions to multi-AI integration and intelligent information systems.
- RELATED WORK
The concept of meta-AI systems draws from the collection of learning in machine learning, where multiple models are combined to improve performance [1]. Early works involved stacking many large-language models. [2] aggregate predictions from diverse classifiers, similar to MetaMind’s ranking approach.
In natural language processing (NLP), systems like IBM Watson [3] and Google BERT LLM[4] integrate multiple models for question-answering. Recent advancements include multi-LLM frameworks, such as Hugging Face’s Transformers library [5], which facilitates model aggregation.
Projects like LangChain [6] enable chaining LLMs for complex tasks, but few focus on real-time comparison and ranking for end-user queries.
Another notable example here would be n8n, which lets the client use LLMs to do complex automation tasks and make useful nodal networks for various computer operations.
Comparative studies, such as those evaluating LLMs on benchmarks like GLUE [7] or MMLU [8], reveal performance variances, underscoring the need for systems like MetaMind. Unlike proprietary platforms (e.g., OpenAI’s API orchestration [9]), MetaMind emphasises open-source integration and a custom ranking engine, making it accessible for educational and research applications.
Framework Multi-LLM support Ranking Mechanism Open Source Focus LangChain Yes (Chaining) No Yes Automation AutoGen Yes (Agents) Partial Yes Collaboration MetaMind Yes (Parallel) Yes (Weighted) Yes Query Reliability - PROBLEM STATEMENT
The main challenge in AI-driven query resolution is the variance across different LLM models. For example, the same query might produce factually accurate but shallow answers from one model, while another provides more depth but introduces biases. This variability comes from: Differing Training Data and Architectures. Models like Gemma and Mistral prioritise efficiency, while DeepSeek and Qwen focus on specialised reasoning.
Reliability Gap: There is no single platform to compare and rank outputs from multiple AIs, which forces users to manually verify responses.
Single-Model Risks.: Relying on just one AI can result in incomplete or incorrect information, which is especially critical in high-stakes situations.
Hallucinations and Factual Inaccuracies: LLMs often generate responses based on various sources. The data available on the internet does not guarantee accuracy, which increases the noise and biases in the LLM.
Bias and Overgeneralization: Imbalances in training data lead to oversimplifications, creating various biases and making the work of model training engineers increasingly difficult. For instance, models may falsely attribute higher risks in certain areas without enough evidence or make broader claims than justified.
Prompt Sensitivity: Small changes in how a query is phrased can lead to significant differences in outputs due to probabilistic sampling. Additionally, models can be affected by tokenisation and preprocessing changes. Users need a smart system that evaluates, compares, and synthesises responses to ensure reliable, trustworthy outputs. Relying on a single LLM for answers is not effective, as the biases and noise in that model can mislead users. Comparing outputs from different models helps confirm the reliability of LLM responses. Current solutions do not scale well for adding new models and do not address latency or cost issues related to multiple API calls.
- OBJECTIVES
The primary objectives of MetaMind are:/p>
Unified Frontend Interface: Develop and implement a responsive, user-friendly and intuitive basic Frontend-Stack search experience with responsive design for desktop and mobile devices.
Backend API Integration: Implement a Node.js server to orchestrate simultaneous requests to Gemma, DeepSeek, Qwen, and potentially Mistral. Node.js allows for simpler callback of the LLM models and also helps in efficient reception of the responses from the LLM model selected.
Intelligent Comparison Engine: Design and implement a custom scoring mechanism that evaluates LLM responses
across four key dimensions: accuracy (fact-checking via semantic similarity or external references), relevance (query alignment using embeddings and cosine similarity), completeness (information depth and coverage scoring), and clarity (readability metrics like Flesch-Kincaid combined with structural analysis).
Optimised Output Delivery: Deliver the highest-ranked answer or a synthesised summary of the best responses. Ensuring the system delivers the best and most efficient answer for the client.
Scalable Architecture: Design a flexible infrastructure to facilitate the easy addition of new AI models. By prioritising open-source compatibility and loose coupling between the frontend and backend, the architecture supports long-term maintenance and adaptation to emerging trends.
Evaluate System Performance: Conduct preliminary testing and diversify query sets to quantify improvements. This objective ensures that the system meets its reliability.
These objectives aim to enhance user experience, reduce errors, and promote multi-AI synergy.
- SYSTEM ARCHITECTURE
MetaMind employs a full-stack architecture to ensure seamless integration and performance:
Frontend Layer
Built with HTML, CSS, and JavaScript for a responsive, engaging user interface.
The current scenario with JavaScript helps in much easier interaction with the frontend and improved usability for the project handlers.
Ensures cross-device compatibility with responsive design principles.
Backend Layer
Node.js with Express.js manages API requests, concurrency, and business logic.
Handles simultaneous queries to multiple AI models via their respective APIs.
Aggregates responses for processing by the ranking engine.
Multi-AI Integration
Queries Gemma, DeepSeek, Qwen (and Mistral) in parallel.
Response aggregation includes temporary storage for comparison.
Ranking System
Applies a weighted scoring mechanism (detailed in Section 6) to evaluate and select/synthesise outputs.
This architecture supports scalability, allowing future expansions to dozens of models with adaptive routing.
- METHODOLOGY
Query Processing Workflow
- User submits a query via the frontend.
- Backend forwards the query to integrated AI models.
- Responses are collected and passed to the ranking engine.
- Ranked or synthesised output is returned to the user.
Answer Ranking Methodology
The ranking engine uses a composite quality metric across four dimensions:
Accuracy Verification: Cross-references facts against reliable sources (e.g., via semantic similarity checks or external knowledge bases).
Relevance Score: Assesses alignment with query intent using NLP techniques like cosine similarity on embeddings.
Completeness Check: Evaluates depth and coverage, scoring based on information density and topic comprehensiveness.
Clarity Rating: Measures readability through metrics like Flesch-Kincaid score, organisation, and logical flow.
Weights are assigned (e.g., 30% accuracy, 25% relevance, 25% completeness, 20% clarity) and normalised to produce a final score. The highest-scoring response is selected, or a hybrid summary is generated if scores are close.
Implementation Details
Software Stack: Windows 10/11, Linux, or macOS; Modern JavaScript for frontend; Node.js for backend; Visual Studio Code as IDE.
Hardware Specifications: Intel i9 13900H processor; 8GB RAM (recommended); 500GB SSD; Integrated graphics; Stable broadband.
AI Models: Integrated via APIs for Gemma, DeepSeek, Qwen, and Mistral.
Initial prototypes used ChatGPT, DeepSeek, and Blackbox for validation, ensuring backward compatibility. This aided in better and efficient development of the MetaMind project.
- USE CASES
MetaMind’s multi-LLM comparative framework uses parallel querying of models like Gemma, DeepSeek, Qwen, and Mistral, followed by smart ranking and synthesis. This provides broad applications across various fields. By combining the strengths of these models, in areas like creative fluency, technical precision, multilingual capability, and efficient inference, the system delivers outputs that are more reliable, nuanced, and contextually suitable than using any single LLM alone. Below, we discuss key real-world applications and show how MetaMind addresses specific problems in each area.
Education and Research: In education, students and teachers often face inconsistent or incomplete AI-generated explanations, especially on complex topics. MetaMind addresses this by creating high-quality, cross-verified responses tailored to what learners need.
Personalised Academic Support: For a question like “Explain Newton’s laws with real-world examples,” MetaMind ranks DeepSeek’s accurate technical breakdown highest for reliability while also using Gemma’s engaging analogies for better clarity. This results in a balanced, student- friendly explanation. It supports self-directed learning, helps with homework, and prepares students for exams, reducing the chances of misinformation that can arise from using a single model.
Intelligent Tutoring and Adaptive Learning: Following trends in LLM-based tutoring, MetaMind works as a virtual tutor by providing step-by-step reasoning, using DeepSeek and Mistral, and offering multilingual explanations through Qwen. In research settings, it assists with literature synthesis, like summarising conflicting papers on climate models. It ranks outputs based on factual completeness and creates hybrid summaries that point out where there is agreement and where there are debates.
Impact: Early user feedback from polytechnic peers shows greater confidence in their answers for assignments and project reports. This aligns with new educational LLM applications, where multi-model approaches boost engagement, lower factual errors, and support different learning styles, such as visual versus analytical learners.
Single LLMs often struggle with multi-step or nuanced reasoning tasks, leading to weak or inconsistent logic. MetaMind takes advantage of model diversity to deliver stronger, more layered reasoning. Complex Analytical Queries: For technical issues, like “Debug this Python sorting algorithm and suggest optimisations,” DeepSeek usually excels in depth of analysis, while Mistral provides efficient alternatives. MetaMind ranks and synthesises these models into a complete answer with checked correctness, different suggestions, and ways to handle edge cases. Multifaceted Decision Support: In situations where balanced viewpoints
are needed, like “Pros and cons of quantum vs. classical computing for AI training,” the system combines creative insights from Gemma, precise comparisons from DeepSeek, and broad knowledge from Qwen. This results in well- rounded outputs that promote critical thinking.
This feature makes MetaMind a valuable resource for advanced reasoning in STEM education, competitive programming, and research hypothesis development.
Evaluation and Results
To assess MetaMinds effectiveness, we conducted preliminary experiments o a dataset of 20 diverse queries by us, covering domains like reasoning, creative responses, and factual knowledge, along with multilingual queries. The queries were sourced from common colleague/research scenarios and benchmarks like subsets of MMLU and TruthfulQA-inspired prompts.
Evaluation Methodology
Responses from MetaMind were recorded mainly on the Gemma 4:B model. The model favours general reasoning and factual knowledge, etc.
- DISCUSSION
Benefits
Enhanced Accuracy: Comparative analysis yields more reliable answers. This enables better accuracy in fields of research and helps increase the certainty of the data.
Improved User Experience: Reduces effort with optimal responses. Saves time on additional research used to cross- check, as multiple LLMs search a single prompt and provide varying responses.
Domain Flexibility: Scalable for tailored AI integrations. Can add more and more open-source LLM models to better evaluate the responses.
Educational Value: Exposes users to diverse perspectives. Aids in better prompt responses and ensures the user receives the correct information.
Challenges and Limitations
Localised Usage: Serverless rendering limits scalability. This impacts the speed with which the response can be received by MetaMinds backend.
Cost Structure: Paid APIs increase expenses. This leads to increased cost. Resulting in cost-cutting due to miscellaneous expenditure in the entire project and development phase.
Latency Impact: Multiple calls extend response times. This results in delayed responses. Also, cold-starting the models becomes delayed.
Ranking Accuracy: Requires ongoing refinement.
As the MetaMind ranking system still needs polishing and work done, the ranking system may have anomalies and abnormalities.
Mitigations include caching mechanisms and open-source model prioritisation.
- CONCLUSION
This paper presented MetaMind, a full-stack comparative answer engine that integrates multiple open-source LLMs (Gemma, DeepSeek, Qwen, Mistral) to address variability in query responses. Through parallel querying, weighted ranking across accuracy, relevance, completeness, and clarity, and optimised output delivery, the system achieves measurable improvements in reliabilityapproximately 12 15% over single-model baselines in preliminary tests.
The work contributes to multi-LLM orchestration by providing an accessible, scalable prototype emphasising educational and research applications. It exemplifies practical full-stack development (React frontend, Node.js backend) alongside intelligent AI aggregation, offering a foundation for more robust query-answering tools in an era of diverse LLMs.
- FUTURE WORK
- DISCUSSION
- USE CASES
Several directions remain for enhancement:
- Integrate advanced routing (e.g., query-response mixed routing or learned routers) to dynamically select models pre-inference, reducing latency and cost.
- Expand evaluation with automated benchmarks (e.g., RouterBench subsets) and LLM judges for objective scoring.
- Add real-time learning for adaptive weights based on user feedback or historical performance.
- Deploy as a cloud service (e.g., Vercel/AWS) with user authentication and query history.
- Explore hybrid synthesis using techniques from recent aggregation frameworks, and support more models (e.g., Llama variants).
- Investigate ethical aspects, such as bias mitigation across ensembles.
These extensions could position MetaMind as a versatile meta-AI platform.
- D. Ferrucci et al., “Building Watson: An Overview of the DeepQA Project,” AI Magazine, vol. 31, no. 3, pp. 59-79, 2010.
- J. Devlin et al., “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” arXiv preprint arXiv:1810.04805, 2018.
- T. Wolf et al., “Transformers: State-of-the-Art Natural Language Processing,” Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 38-45, 2020.
- LangChain Documentation, Available: https://langchain.com/docs.
- A. Wang et al., “GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding,” arXiv preprint arXiv:1804.07461, 2018.
- D. Hendrycks et al., “Measuring Massive Multitask Language Understanding,” arXiv preprint arXiv:2009.03300, 2020.
- OpenAI, “ChatGPT API Documentation,” Available: https://openai.com/api. [10] DeepSeek, “DeepSeek AI Platform,” Available: https://deepseek.com.
- Blackbox AI, “Blackbox AI Tools and API,” Available: https://blackbox.ai. [12] React Team, “React Documentation,” Meta Platforms, Available: https://reactjs.org/docs.
- Node.js Foundation, “Node.js Documentation,” Available: https://nodejs.org/docs.
REFERENCES
- L. Breiman, “Stacked Regressions,” Machine Learning, vol. 24, no. 1,
pp. 49-64, 1996.
- D. H. Wolpert, “Stacked Generalisation,” Neural Networks, vol. 5, no. 2, pp. 241-259, 1992.
