Awesome
Mini-Gemini Chatbot Angular client using Gemini API (via API key and Google AI Studio)
Features
- Support for Google AI and VertexAI APIs
- Gemini API (generateContent, startChat and sendMessage)
- Demonstration of Gemini for Text with text-to-speech (ElevenLabs)
- Demonstration of Gemini for Chat with Rich Media support (markdown, code, emojis, formulas and diagrams)
- Demonstration of Gemini for Images with Rich Media support (markdown, code, emojis, formulas and diagrams)
Libraries
- Angular Material
- Gemini for Text:
- ngx-quill - Angular QuillJS wrapper, a Rich Text Editor
- QuillJS. Custom non-editable block (embed blot) to handle model responses.
- ElevenLabs API - High quality natural voices
- Gemini for Chat:
- ngx-markdown - Markdown renderer with support for multiple extensions (code fences and highlighting, emojis, Katex mathematic formulas, MermaidJS diagrams and more)
- Gemini for Images:
- ngx-markdown - Markdown renderer with support for multiple extensions (code fences and highlighting, emojis, Katex mathematic formulas, MermaidJS diagrams and more)
This project was generated with Angular CLI.
Project Set-up
Run this Angular CLI command to get the local environment setup:
ng g environments
This will create the following files for development
and production
:
src/environments/environment.development.ts
src/environments/environment.ts
Change these files to include your environment setup accordingly. For the examples below we will use the development
file:
// src/environments/environment.development.ts
export const environment = {
/// your setup
};
This project includes clients to both VertexAI and Gemini APIs.
Setup for Gemini API access via API key (Google AI Studio)
Get an API key from Google AI Studio, then configure it here.
Note that access is restricted to the US. Use a VPN to overcome this limitation while outside the US.
// src/environments/environment.development.ts
export const environment = {
API_KEY: "<<YOUR-API-KEY-FROM-GOOGLE-AI-STUDIO>>",
};
Setup for API access via VertexAI (Google Cloud)
This setup requires a Google Cloud account
and is more advanced both in security and capabilities. VertexAI has access to more advanced Generative AI features (image, code, voice and more) and foundational models.
To secure the API access, you need to create an account and get the credentials for your application so only you can access it. Here are the steps:
- Sign up for a Google Cloud account and enable billing — this gives you access to Vertex AI.
- Create a new project in the Cloud Console. Make note of the project ID.
- Enable the Vertex AI API for your project.
- Install the gcloud CLI and run gcloud auth print-access-token. Save the printed access token - you’ll use this for authentication.
Once you have the project ID and access token, you are ready to move on to the Angular app. To verify everything is setup correctly you can try these curl commands.
// src/environments/environment.development.ts
export const environment = {
PROJECT_ID: "<<YOUR-PROJECT-ID>>",
GCLOUD_AUTH_PRINT_ACCESS_TOKEN: "<<YOUR-GCLOUD-AUTH-PRINT-ACCESS-TOKEN>>",
};
Make sure that your access token remains private so you do not incur into expenses from unauthorised access.
You can find more details into how to setup the VertexAI access in this article
ElevenLabs Setup
This project supports the regular
and streamed
APIs for faster responsiveness.
You can use the newest eleven_multilingual_v2
, a single foundational model supporting 28 languages including English, Chinese, Spanish, Hindi, Portuguese, French, German, Japanese, Arabic, Korean, Indonesian, Italian, Dutch, Turkish, Polish, Swedish, Filipino, Malay, Romanian, Ukrainian, Greek, Czech, Danish, Finnish, Bulgarian, Croatian, Slovak, and Tamil; or eleven_monolingual_v1
, a low-latency model specifically trained for English speech.
You need to generate your API key (Sign In/Profile).
Picking a voice
Using your API key you can either use Speech Synthesis to listen to the voices and open Developer Tools/Network to extract the voice-id from the request or use the API as shown below:
curl -X 'GET' \
'https://api.elevenlabs.io/v1/voices' \
-H 'accept: application/json' \
-H 'xi-api-key: <<API-KEY>>'
Pick a voice from the response
{
"voices": [
{
"voice_id": "21m00Tcm4TlvDq8ikWAM",
"name": "Rachel",
"category": "premade",
"labels": {
"accent": "american",
"description": "calm",
"age": "young",
"gender": "female",
"use case": "narration"
},
"preview_url": "https://storage.googleapis.com/eleven-public-prod/premade/voices/21m00Tcm4TlvDq8ikWAM/df6788f9-5c96-470d-8312-aab3b3d8f50a.mp3",
...
},
]
}
Once you have the API key
and voice-id
you can fill in the remaining entries in the environment file.
// src/environments/environment.development.ts
export const environment = {
ELEVEN_LABS_API_KEY: "<<ELEVEN-LABS-API-KEY>>",
ELEVEN_LABS_VOICE_ID: "<<ELEVEN-LABS-VOICE-ID>>",
};
Development server
Run ng serve
for a dev server. Navigate to http://localhost:4200/
. The application will automatically reload if you change any of the source files.
Code scaffolding
Run ng generate component component-name
to generate a new component. You can also use ng generate directive|pipe|service|class|guard|interface|enum|module
.
Build
Run ng build
to build the project. The build artifacts will be stored in the dist/
directory.
Running unit tests
Run ng test
to execute the unit tests via Karma.
Running end-to-end tests
Run ng e2e
to execute the end-to-end tests via a platform of your choice. To use this command, you need to first add a package that implements end-to-end testing capabilities.
Further help
To get more help on the Angular CLI use ng help
or go check out the Angular CLI Overview and Command Reference page.