Efficient Strategy for Improving Large Language Model (LLM) Capabilities
By: Julián Camilo Velandia Gutiérrez
Potential Business Impact:
Makes smart computer programs run faster with less power.
Large Language Models (LLMs) have become a milestone in the field of artificial intelligence and natural language processing. However, their large-scale deployment remains constrained by the need for significant computational resources. This work proposes starting from a base model to explore and combine data processing and careful data selection techniques, training strategies, and architectural adjustments to improve the efficiency of LLMs in resource-constrained environments and within a delimited knowledge base. The methodological approach included defining criteria for building reliable datasets, conducting controlled experiments with different configurations, and systematically evaluating the resulting variants in terms of capability, versatility, response time, and safety. Finally, comparative tests were conducted to measure the performance of the developed variants and to validate the effectiveness of the proposed strategies. This work is based on the master's thesis in Systems and Computer Engineering titled "Efficient Strategy for Improving the Capabilities of Large Language Models (LLMs)".
Similar Papers
Strategic Innovation Management in the Age of Large Language Models Market Intelligence, Adaptive R&D, and Ethical Governance
Computation and Language
Speeds up inventing new things by helping scientists.
Large Language Models Meet Legal Artificial Intelligence: A Survey
Computation and Language
Helps lawyers use smart computers for legal work.
Method-Based Reasoning for Large Language Models: Extraction, Reuse, and Continuous Improvement
Computational Engineering, Finance, and Science
Teaches computers to solve new problems logically.