Building An AI Agent Based On CAN-20 Protocol Primitive Data

This chapter focuses on training an AI agent using the initial CAN-20 protocol's generated data.

Introduction

Leveraging the General Language Model (GLM) architecture and model quantization, we fine-tune using the high-efficiency P-Tuning v2 method alongside user-uploaded on-chain data. This produces Chain-GLM — a cost-effective web-enhanced question answering system similar to WebGLM. Chain-GLM is designed to utilize decentralized on-chain data by integrating search/retrieval into pre-trained language models. This allows freely distributing information and accurately obtaining it via AI capabilities.

Implementation

ChainGLM aims to improve real-world application deployment by integrating web search and retrieval capabilities into pre-trained language models.

  • LLM Enhanced Retriever: Retrieves relevant web content to answer questions more accurately

  • Elicitation Generator: Leverages GLM to generate high-quality, human-like responses

  • Human Preference-Aware Scorer: Prioritizes useful, engaging responses based on human preferences.

  • P-Tuning v2

    • P-Tuning v2 is an optimized cue tuning strategy that achieves performance comparable to fine-tuning on small/medium models and sequence labeling challenges.

    • P-tuning v2 utilizes deep cue tuning, where successive cues are applied to each layer input of the pretrained transformer. Deep cue tuning increases the ability to continuously cue and closes the gap with fine-tuning in a variety of settings, especially for small models and difficult tasks.

  • Integration with on-chain data

Last updated