Fine-Tuning Gemma-2B for Text-to-SQL Generation
Through this study, the concept of large language models (LLM) in the context of generative AI has been explored. Gemma-2B, with 4-bit quantization is fine-tuned to bridge the gap between natural language processing (NLP) and SQL query generation. The 4-bit quantization significantly reduces memory...
Saved in:
Published in | 2024 9th International Conference on Communication and Electronics Systems (ICCES) pp. 265 - 269 |
---|---|
Main Authors | , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
16.12.2024
|
Subjects | |
Online Access | Get full text |
DOI | 10.1109/ICCES63552.2024.10860111 |
Cover
Summary: | Through this study, the concept of large language models (LLM) in the context of generative AI has been explored. Gemma-2B, with 4-bit quantization is fine-tuned to bridge the gap between natural language processing (NLP) and SQL query generation. The 4-bit quantization significantly reduces memory requirements, making Gemma-2B more efficient for real-world applications while maintaining the performance of the large language model. Users can interact with databases intuitively through simple, text-based inputs. This model employs zero-shot learning, enabling SQL query generation for text descriptions without specific examples in training. Developed and tested within a Google Colab environment, this approach simplifies data retrieval by transforming unstructured language into structured queries through generative AI. |
---|---|
DOI: | 10.1109/ICCES63552.2024.10860111 |