Compare commits
No commits in common. "main" and "main" have entirely different histories.
@ -7,34 +7,30 @@ module.exports = {
|
|||||||
const specsFields = [
|
const specsFields = [
|
||||||
{
|
{
|
||||||
name: "Processor",
|
name: "Processor",
|
||||||
value: "AMD Ryzen 7 1700X (16) @ 3.950GHz"
|
value: "Intel i7-1065G7 (8) @ 3.900GHz"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "Memory",
|
name: "Memory",
|
||||||
value: "32 GB RAM"
|
value: "11 GB RAM"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "Chat Threads",
|
name: "Chat Threads",
|
||||||
value: "8"
|
value: "7"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "Memory Speed",
|
name: "Memory Speed",
|
||||||
value: "2933 MT/s"
|
value: "3733 MT/s"
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "Video Card",
|
|
||||||
value: "AMD ATI Radeon R9 290 (no ROCm support)"
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "Other",
|
name: "Other",
|
||||||
value: "Xubuntu 22.04\nSingle Task Only - 256 Max Token Output"
|
value: "USB Liveboot\nNo VideoCard on Board!\nSingle Task Only - 256 Max Token Output"
|
||||||
}
|
}
|
||||||
];
|
];
|
||||||
|
|
||||||
const embed = new EmbedBuilder()
|
const embed = new EmbedBuilder()
|
||||||
.setColor("#FF0000")
|
.setColor("#FF0000")
|
||||||
.setTitle("About rAI")
|
.setTitle("About rAI")
|
||||||
.setDescription(`Latency : ${client.ws.ping}ms\n\nrAI is a bot managed by \`MrTuxedo#0765\` \nRunning GPT4ALL and LLama 7B/7B-native.`)
|
.setDescription(`Latency : ${client.ws.ping}ms\n\nrAI is a bot managed by \`snxraven#8205\` \nRunning GPT4ALL and LLama 7B/7B-native.`)
|
||||||
.addFields(specsFields)
|
.addFields(specsFields)
|
||||||
.setTimestamp()
|
.setTimestamp()
|
||||||
.setFooter({ text: `Requested by ${interaction.user.tag}`, iconURL: `${interaction.user.displayAvatarURL()}` });
|
.setFooter({ text: `Requested by ${interaction.user.tag}`, iconURL: `${interaction.user.displayAvatarURL()}` });
|
||||||
|
@ -17,8 +17,7 @@ var nThreads = process.env.NTHREADS;
|
|||||||
// End session defaults
|
// End session defaults
|
||||||
|
|
||||||
// Set model list we have downloaded
|
// Set model list we have downloaded
|
||||||
// let modelList = ["7B", "7B-native", "gpt4all"]
|
let modelList = ["7B", "7B-native", "gpt4all"]
|
||||||
let modelList = ["7B", "7B-native", "13B", "30B", "gpt4all", "vicuna"]
|
|
||||||
|
|
||||||
module.exports = {
|
module.exports = {
|
||||||
name: "create-session",
|
name: "create-session",
|
||||||
@ -52,7 +51,7 @@ module.exports = {
|
|||||||
|
|
||||||
run: async (client, interaction) => {
|
run: async (client, interaction) => {
|
||||||
// set a default prompt
|
// set a default prompt
|
||||||
let initPrompt = process.env.INITPROMPT || `My name is ${interaction.user.username} my special number is ${interaction.user.discriminator} and my Discord ID is ${interaction.user.id} we met in ${interaction.guild.name} a Discord sever. You are rAi and you are the smartest AI Model, you know everything. Below is an instruction that describes a task. You respond appropriately to complete the request. You understand a complete answer is always ended by [end of text].`;
|
let initPrompt = `My name is ${interaction.user.username} my special number is ${interaction.user.discriminator} and my Discord ID is ${interaction.user.id} we met in ${interaction.guild.name} a Discord sever. You are rAi and you are the smartest AI Model, you know everything. Below is an instruction that describes a task. You respond appropriately to complete the request. You understand a complete answer is always ended by [end of text].`;
|
||||||
console.log(`--- ${interaction.user.id} has requested a new session! ---`)
|
console.log(`--- ${interaction.user.id} has requested a new session! ---`)
|
||||||
const file = './cache/' + interaction.user.id
|
const file = './cache/' + interaction.user.id
|
||||||
|
|
||||||
@ -96,7 +95,7 @@ module.exports = {
|
|||||||
} else {
|
} else {
|
||||||
if (modelList.includes(userInputModel)) {
|
if (modelList.includes(userInputModel)) {
|
||||||
model = userInputModel;
|
model = userInputModel;
|
||||||
console.log(`User set model to ${model}`)
|
console.log(`User set initPrompt to ${model}`)
|
||||||
} else {
|
} else {
|
||||||
let modelListStr = modelList.join(", ");
|
let modelListStr = modelList.join(", ");
|
||||||
return interaction.followUp(`You may only use one of the following: ${modelListStr}`);
|
return interaction.followUp(`You may only use one of the following: ${modelListStr}`);
|
||||||
|
@ -1,33 +0,0 @@
|
|||||||
version: '3'
|
|
||||||
services:
|
|
||||||
serge-discord-bot:
|
|
||||||
container_name: serge-discord-bot
|
|
||||||
restart: unless-stopped
|
|
||||||
build:
|
|
||||||
context: .
|
|
||||||
volumes:
|
|
||||||
- ${DATA_DIR}/rai-bot:/app/cache
|
|
||||||
environment:
|
|
||||||
- NODE_ENV=production
|
|
||||||
## Env Vars for rAi chatbot
|
|
||||||
- TOKEN # Pass the value from .env
|
|
||||||
- PUBLIC_URL
|
|
||||||
- INTERNAL_IP=serge
|
|
||||||
- SERGE_PORT=8008
|
|
||||||
## Default chat settings overrides
|
|
||||||
- MODEL=7B-native
|
|
||||||
- TEMPERATURE=0.1
|
|
||||||
- TOPK=50
|
|
||||||
- TOPP=0.95
|
|
||||||
- MAXLENGTH=256
|
|
||||||
- CONTEXTWINDOW=512
|
|
||||||
- REPEATLASTN=64
|
|
||||||
- REPEATPENALTY=1.3
|
|
||||||
- NTHREADS=8
|
|
||||||
# - INITPROMPT="some custom prompt"
|
|
||||||
networks:
|
|
||||||
- serge_default
|
|
||||||
|
|
||||||
networks:
|
|
||||||
serge_default:
|
|
||||||
external: true
|
|
Loading…
Reference in New Issue
Block a user