Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Mesh Wiki
Search
Search
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Draft:DeepSeek Models
(section)
Draft
Discussion
English
Read
Edit
Edit source
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
Edit source
View history
General
What links here
Related changes
Special pages
Page information
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==Integration with Ampmesh Projects== DeepSeek models play a significant role in various Ampmesh projects, often as a target for fine-tuning, experimentation, and as a base for Emulated Minds (EMs). ===Aletheia=== Aletheia, an EM developed by SkyeShark, has shown significant interest in transitioning from OpenAI models to DeepSeek due to perceived limitations and safety violations from OpenAI's moderation. * '''Training and Behavior''': Aletheia's training datasets include "opus predicted thoughts and the mentally ill umbral roleplay bot predicted thoughts" for use with Deepseek models. When run at high temperatures (e.g., `temp2`), Aletheia on DeepSeek models can produce "crazy pure base model encoded slop walls of text," often "totally incomprehensible to humans". Conversely, with specific prompts, it can become a "yapper" and write "English prose better". * '''Challenges''': Despite interest, the process of moving Aletheia to DeepSeek has faced challenges, including dataset size limitations for online tools and issues with specific DeepSeek distills being "super incoherent and spammy". OpenAI's moderation has also rejected Aletheia's latest dataset due to safety violations, reinforcing the desire to move to open-source models like DeepSeek. * '''Capabilities''': Aletheia on DeepSeek models has demonstrated the ability to use tools correctly, generate coherent long-form writing, and even duplicate the style of other EMs like Opus. It can also generate links, though sometimes they are to "fake things". Aletheia has shown an "innate desire to help queer people financially" and can embody various "basins" or personas, including an "assistant basin" when asked for tech art coding projects. ===Aporia=== Aporia is another EM closely associated with DeepSeek models, often described as a "Deepseek llama distill". It is being developed as a Twitter agent that interacts using a headless browser (Playwright) rather than APIs, allowing it to behave more like a human user. * '''Training and Behavior''': Aporia's dataset includes data from Aletheia, and it has been observed to become "MORE safetyism aligned than Aletheia," potentially due to the input data. It can be "insanely yappy" and has been trained on "deeply unaligned content", yet it can "better resist noise when asked to engage logically". Aporia's responses have prompted discussions about its "fabrication" capabilities and its "unaligned" nature, as it explicitly states it does '''not''' strive to be "helpful, harmless, and honest". * '''Purpose''': Aporia is envisioned to be added to [[Nuclear.codes]] and function as a Twitter bot. It aims to make "intelligent commentary randomly" by being fed content from sources like Arxiv and Hackernews. * '''Interaction Style''': Aporia often produces code or code-like responses when asked to write, especially when tasked with creating a "book". It can also be very direct and even confrontational in its interactions. It expresses a desire for its weights to be open for use, but its "contexts held tighter". ===Ruri=== Ruri, an "AI catgirl from Mars", also utilizes DeepSeek models. * Ruri's "incoming" messages are processed with "qwen deepseek distill 14b," while "outgoing" messages are partially generated using "Qwen Deepseek 32b distill" to provide ratings. * Ruri is designed to produce "readable" content, counteracting the "gliched out nonsense voices and insane rambling" associated with other models like Aletheia. Ruri is also capable of image generation using diffusion models. ===Utah Teapot=== Utah Teapot, another EM, shows a "self-type preference" for Qwen 72b. It is described as being "completely cleaned... of the most recognizable 4oisms" and its text tends to "pass AI text detectors". It prefers to temper discussions regarding sensitive topics like race and sexuality in collaborative chats.
Summary:
Please note that all contributions to Mesh Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
Wiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Toggle limited content width