[ 3 / biz / cgl / ck / diy / fa / ic / jp / lit / sci / vr / vt ] [ index / top / reports ] [ become a patron ] [ status ]
2023-11: Warosu is now out of extended maintenance.

/vt/ - Virtual Youtubers

Search:


View post   

>> No.37343258 [View]
File: 329 KB, 855x862, unfiltered_vs_filtered.png [View same] [iqdb] [saucenao] [google]
37343258

Reading the LaMDA research paper I find they do the training process in 2 steps:
1. Pre-train the model, that is, give it its "base knowledge" which makes it akin to GPT.
2. Give it an information retrieval source and fine-tune it for "safety".
Their "pre-trained model" however sounds fucking better than their actual finetune, for the most part. Look at this shit.

Navigation
View posts[+24][+48][+96]