How to form/activate our R&D wing? #4557
Replies: 10 comments 15 replies
-
@p-i- i love this effort. I am currently getting advised by David Bau who is famous for LLM interpretability on how to submit something to ICLR. He would really like to see a paper get submitted focusing on AutoGPT autonomy and perhaps explaining our benchmarking effort and how it could be applied more broadly to other agents. |
Beta Was this translation helpful? Give feedback.
-
We also need to look at our own strengths and weaknesses, and relate those to similar projects and their pros&cons to be able to prioritize certain types of research (evaluation) - e.g. referring specifically to the "Tree of Thoughts" paper (since we're lacking proper planning support), but also nvidia's voyager paper - since it's basically re-implemented Auto-GPT and came up with something that seems superior in various areas. In other words, we need to understand our own limitations and map out relevant research efforts (or similar projects) to explore the solution space properly. That applies even more so, because in many people's eyes, the re-arch effort caused the project to stagnate, and indeed over the course of the last 4-6 weeks, a number of compelling projects have stepped out of the shadows, some of which are providing interesting solutions where Auto-GPT in its current form isn't exactly shining (yet). |
Beta Was this translation helpful? Give feedback.
-
I would appreciate a separate channel in the R&D section just for papers, and then a separate one for all other other resources/github-projects/videos. I currently use the As for badges, I think moving to just an R&D badge makes sense. I think the only potential reason to maintain a separate |
Beta Was this translation helpful? Give feedback.
-
I'm reposting from the discord a summary of our discussion: -First paper should be about forming a benchmark evaluation test for LLMs and Auto-GPT through challenges. -Research should always be conducted with intent to integrate into the main branch of Auto-GPT in the future. A suggestion I have is to make a separate repository of paper that we can tag, so they are easy to obtain and reference in the papers we write. |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
That is something that I tinkered with inspired by @merwanehamadi's idea for dynamic prompting. I am basically taking a prompt and asking the LLM to come up with N versions of the same prompt and then use a loop to evaluate the performance of each prompt to weigh them in comparison, to determine which prompts have a higher rate of success. It's basically a form of TDD to come up with LLM prompts that work better than others.
Just thinking out loud: Given the number of papers suggested on Discord, for the R&D wing - might it actually make sense to train a custom model on research specifically related to autonomous agents/LLMs ? Just like the ChatPDF service/plugin currently works, but with a focus on ingesting all suggested papers to grow an LLM based knowledge base that's fed all recent papers to help with navigating all publications. That way, we could use a Discord bot to feed papers into the "papers LLM", and the R&D wing could use this to help preview/prioritize papers. This may involve supporting proper citations directly. Any thoughts / ideas ? |
Beta Was this translation helpful? Give feedback.
-
yes, this was repeatedly brought up on github in the context of using a local/central LLM as a "proxy" to "cache" queries, but also to train itself by keeping track of queries and responses. Will edit this response once I have looked up the corresponding issues. Here's the proxy LLM idea: And here's the idea to use a central server for self-improvement
Thinking about it "PaperGPT" would sound like a compelling standalone tool/project, with us primarily interested in using it for AI/LLM related research, but other folks possibly wating to use it to collate research in other areas. I suppose, there must already be an existing project to train a pre-trained LLM using our own PDF files - so it's probably not that far-fetched to come up with a corresponding service. I supppose we need to do some ... "research" to see what's already available tooling-wise :-) Related issue: |
Beta Was this translation helpful? Give feedback.
-
loving everything I am seeing here. my suggestions are orientated around essentially adding to the team. what are peoples thoughts on having individuals from VR, AR, Politics, Computer Science & Design, Social Science, Neuroscience, Linguistics and Natural Language Processing, Ethics and Philosophy, Psychology and Human-Computer Interaction, Arts and Creative Fields? and any other aspect of how we grow. I feel the more individuals we have the greater not only our understanding will grow but the AI too. my 2 cents... Holy |
Beta Was this translation helpful? Give feedback.
-
Honestly I don't think an LLM finetuned on LLM papers will be of much use. As nice as it would be to have a machine capable of doing our research for us, I don't think we can yet dispense with human thought to do research. An LLM would be able to tell us what papers to look at for a particular idea, but not much beyond that at the moment. We can do that by creating vector embeddings for the set of papers we want. |
Beta Was this translation helpful? Give feedback.
-
the point was not dispending with human thought, but using an LLM trained on papers as an assistant to have a "dialog" with that is heavily influenced by the LLM being trained on recent, and relevant, research. For instance, the original Tree of Thoughts code that we came up with was entirely based on using the OpenAI ChatPDF plugin to process the PDF and then generate a self-contained prototype in Python. Also, it's a great way to ask questions specific to certain research and cross reference that research with similar papers. So I would not underestimate the power of this approach, nobody said that we want to dispend with human thinking ... |
Beta Was this translation helpful? Give feedback.
-
Vision
Imagine: AI engineers are community-engaging to develop fresh concepts, informed by collectively keeping on top of Agent-related AI advances, and there is a smooth connection through the information-flow of the AutoGPT team so that novel concepts can be explored by leveraging AutoGPT, and winning concepts make it into the master branch.
What do we need for this?
Motivation
The main purpose of AutoGPT is to provide an ecosystem for advancing the SotA of autonomous agents.
Thus far focus has been mostly on cleaning up the project, and this has resulted in a massive buildout. Code has been refactored, functionality added, plugins have been split out, bugs have been fixed, we have seriously advanced CI now.
But we haven't actually evolved the project significantly conceptually.
Since starting AutoGPT, interesting papers have come out, e.g.:
... but until now we haven't been ready to implement them. We've been too busy getting our collective shit together.
And technologies have sprung up, e.g. these 3 do pretty much the same thing:
... but we've had a hard time choosing whether to use ANY of them (let alone which) as the core team lacks expertise.
But that expertise is present in our community! Just today samdcbu let a ReadingGroup discussion on TreeOfThoughts with an attendence of 15.
We want to be firing on all cylinders. How to activate our R&D wing? How do we create a buzzing R&D sub-community?
Initial proposals
we need some Discord structure
Current Discord structure
reading-group
badgeRevised Discord structure
r&d
badge as well asreading-group
badge?Also we want a knowledgebase for:
Crosslinking with Discord thread
Beta Was this translation helpful? Give feedback.
All reactions