Files
encoach_frontend/src/pages/api/evaluate/interactiveSpeaking.ts
2023-09-19 00:57:36 +01:00

60 lines
1.7 KiB
TypeScript

// Next.js API route support: https://nextjs.org/docs/api-routes/introduction
import type {NextApiRequest, NextApiResponse} from "next";
import {withIronSessionApiRoute} from "iron-session/next";
import {sessionOptions} from "@/lib/session";
import axios from "axios";
import formidable from "formidable-serverless";
import {getStorage, ref, uploadBytes} from "firebase/storage";
import fs from "fs";
import {app} from "@/firebase";
export default withIronSessionApiRoute(handler, sessionOptions);
async function handler(req: NextApiRequest, res: NextApiResponse) {
if (!req.session.user) {
res.status(401).json({ok: false});
return;
}
const storage = getStorage(app);
const form = formidable({keepExtensions: true});
await form.parse(req, async (err: any, fields: any, files: any) => {
if (err) console.log(err);
const uploadingAudios = await Promise.all(
Object.keys(files).map(async (fileID: string) => {
const audioFile = files[fileID];
const questionID = fileID.replace("answer_", "question_");
const audioFileRef = ref(storage, `speaking_recordings/${(audioFile as any).path.replace("upload_", "")}`);
const binary = fs.readFileSync((audioFile as any).path).buffer;
const snapshot = await uploadBytes(audioFileRef, binary);
fs.rmSync((audioFile as any).path);
return {question: fields[questionID], answer: snapshot.metadata.fullPath};
}),
);
const backendRequest = await axios.post(
`${process.env.BACKEND_URL}/speaking_task_3`,
{answers: uploadingAudios},
{
headers: {
Authorization: `Bearer ${process.env.BACKEND_JWT}`,
},
},
);
res.status(200).json({...backendRequest.data, answer: uploadingAudios});
});
}
export const config = {
api: {
bodyParser: false,
},
};