Compare commits

..

57 Commits

Author SHA1 Message Date
Jonathan Dunn
a6aeb8ffed added agents 2024-02-28 10:17:57 -05:00
Daniel Miessler
1c71ac790d Updated rpg_summarizer. 2024-02-25 11:13:12 -06:00
Daniel Miessler
c15d043bc6 Updated rpg_summarizer. 2024-02-25 11:08:10 -06:00
jad2121
7c1b819ffc fixed more stuff 2024-02-24 16:49:45 -05:00
jad2121
ea7460d190 fixed something 2024-02-24 16:39:48 -05:00
Daniel Miessler
e8c8ea10dc Updated README.md with video info. 2024-02-23 20:50:26 -08:00
Daniel Miessler
4146460c76 Updated README.md with video info. 2024-02-23 20:47:37 -08:00
Daniel Miessler
bb57e4a241 Updated README.md with video info. 2024-02-23 20:43:10 -08:00
Daniel Miessler
5e56731032 Updated README.md with video info. 2024-02-23 20:42:14 -08:00
Daniel Miessler
8aa88909a8 Updated README.md with video info. 2024-02-23 20:39:58 -08:00
Daniel Miessler
aff74ec628 Updated README.md with video info. 2024-02-23 20:37:12 -08:00
Daniel Miessler
f1cfaf0ed3 Updated README.md with video info. 2024-02-23 20:33:56 -08:00
Daniel Miessler
8f90b8db06 Updated README.md with video info. 2024-02-23 20:30:11 -08:00
Daniel Miessler
3c32e3266d Updated README.md with video info. 2024-02-23 20:29:18 -08:00
Daniel Miessler
f73299d999 Updated README.md with video info. 2024-02-23 20:27:19 -08:00
Daniel Miessler
90f96b0f37 Updated README.md with video info. 2024-02-23 20:25:00 -08:00
Daniel Miessler
4377838822 Updated README.md with video info. 2024-02-23 20:24:15 -08:00
Daniel Miessler
d1a8976a64 Updated intro video. 2024-02-23 20:22:01 -08:00
Daniel Miessler
d64434e8ca Merge pull request #125 from danielmiessler/dependabot/pip/cryptography-42.0.4
Bump the pip group across 1 directories with 1 update
2024-02-23 20:08:49 -08:00
Daniel Miessler
25de07504c Merge pull request #129 from arduino-man/main
Alphabetically sort patterns list
2024-02-23 13:33:04 -08:00
Daniel Miessler
524393ba7d Updated readme for server instructions. 2024-02-23 13:26:14 -08:00
Daniel Miessler
d129188da8 Updated create_video_chapters. 2024-02-22 16:22:54 -08:00
Daniel Miessler
99e4723a6d Updated create_video_chapters. 2024-02-22 16:19:57 -08:00
Daniel Miessler
2a5646d92f Updated create_video_chapters. 2024-02-22 16:17:19 -08:00
Daniel Miessler
7aba85856c Updated create_video_chapters. 2024-02-22 16:11:01 -08:00
Daniel Miessler
fe5e4ba048 Added create_video_chapters. 2024-02-22 16:06:00 -08:00
Daniel Miessler
729f12917b Updated label_and_rate. 2024-02-21 22:35:32 -08:00
Daniel Miessler
46a58866f4 Updated label_and_rate. 2024-02-21 22:03:11 -08:00
Daniel Miessler
c12bbed32c Updated label_and_rate. 2024-02-21 21:53:50 -08:00
arduino-man
e5901b9f44 Alphabetically sort patterns list
Ensures that when the users lists the available patterns, they are presented in alphabetical order. Helps find the desired pattern faster.
2024-02-21 20:01:22 -07:00
dependabot[bot]
e5e19d7937 Bump the pip group across 1 directories with 1 update
Bumps the pip group with 1 update in the /. directory: [cryptography](https://github.com/pyca/cryptography).


Updates `cryptography` from 42.0.2 to 42.0.4
- [Changelog](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst)
- [Commits](https://github.com/pyca/cryptography/compare/42.0.2...42.0.4)

---
updated-dependencies:
- dependency-name: cryptography
  dependency-type: indirect
...

Signed-off-by: dependabot[bot] <support@github.com>
2024-02-21 20:44:42 +00:00
Daniel Miessler
92f8e08aac Cleanup. 2024-02-21 09:38:07 -08:00
Daniel Miessler
62f3608144 Updated output instructions. 2024-02-21 09:14:17 -08:00
Daniel Miessler
20c1ad90bb Created a STATISTICS version of analyze_threat_report. 2024-02-21 09:11:20 -08:00
Daniel Miessler
e866eeafa6 Created a STATISTICS version of analyze_threat_report. 2024-02-21 09:09:12 -08:00
Daniel Miessler
5e48c0ef2c Created a TRENDS version of analyze_threat_report. 2024-02-21 09:06:02 -08:00
Daniel Miessler
61421c28cb Improved summary to analyze_threat_report. 2024-02-21 09:03:44 -08:00
Daniel Miessler
7ebf5bc905 Added summary to analyze_threat_report. 2024-02-21 09:01:49 -08:00
Daniel Miessler
9cd15d725c Added a threat report analysis pattern. 2024-02-21 08:59:45 -08:00
Jonathan Dunn
138c779f5e changed readme 2024-02-21 08:39:31 -05:00
jad2121
31ab369e2f changed another message 2024-02-21 06:25:21 -05:00
jad2121
983084e4f0 added a statement 2024-02-21 06:24:01 -05:00
jad2121
ed847fd332 Added aliases for individual patterns. Also fixed pattern download process 2024-02-21 06:19:54 -05:00
Daniel Miessler
373d362d35 Merge pull request #118 from mikeprivette/main
Enhanced Setup Script Compatibility and Reliability Improvements
2024-02-20 09:22:28 -08:00
Mike Privette
6dff639969 Updates
- README.md - added instructions to make sure the setup.sh script was executable as this was not explicitly stated

- setup.sh - updated sed to use `sed -i` to be compatible with Linux, MacOSX and other OS versions and added a check in the local directory taht setup.sh executes in for a pyproject.toml file because the script was looking for the .toml file in the user's home directory and throwing an error
2024-02-20 10:41:34 +00:00
Daniel Miessler
6414c26636 Updated write_essay to be more conversational and less grandiose and pompous. 2024-02-19 17:34:22 -08:00
Daniel Miessler
bc4456b310 Merge pull request #114 from fureigh/remove-ds-store
Removes stray .DS_Store file
2024-02-18 18:47:34 -08:00
Daniel Miessler
873bca5230 Merge pull request #115 from fureigh/gerunds-ahoy
Makes a minor README edit for the sake of consistency
2024-02-18 18:47:05 -08:00
Fureigh
5d984f3687 Minor README edit for verb form consistency
Change `Create` to `Creating`.
2024-02-18 17:40:08 -08:00
Fureigh
9863573ff6 Remove stray .DS_Store file 2024-02-18 17:05:08 -08:00
jad2121
335fea353b now context.md is in .config 2024-02-18 16:48:47 -05:00
jad2121
a0d264bead updated readme 2024-02-18 16:36:17 -05:00
jad2121
d15e022abf fixed context 2024-02-18 16:34:47 -05:00
jad2121
8f4ab672c6 added context to cli. edit context.md and add -C to add context to your queries 2024-02-18 13:25:07 -05:00
Daniel Miessler
b127fbec15 Updated analyze_paper with more detail and legibility. 2024-02-17 19:38:12 -08:00
Daniel Miessler
0deab1ebb3 Updated analyze_paper with more detail and legibility. 2024-02-17 19:35:09 -08:00
Daniel Miessler
8aacaee643 Added a specific version of extract_wisdom just for articles. 2024-02-17 19:03:45 -08:00
35 changed files with 2929 additions and 251 deletions

1
.python-version Normal file
View File

@@ -0,0 +1 @@
3.10

View File

@@ -14,6 +14,7 @@
<h4><code>fabric</code> is an open-source framework for augmenting humans using AI.</h4>
</p>
[Introduction Video](#introduction-video) •
[What and Why](#whatandwhy) •
[Philosophy](#philosophy) •
[Quickstart](#quickstart) •
@@ -25,6 +26,7 @@
## Navigation
- [Introduction Video](#introduction-video)
- [What and Why](#what-and-why)
- [Philosophy](#philosophy)
- [Breaking problems into components](#breaking-problems-into-components)
@@ -45,18 +47,13 @@
<br />
> [!NOTE]
> February 16, 2024 — **It's now far easier to install and use Fabric!** Just head to the [Quickstart](#quickstart), install Poetry, and run `setup.sh`, and it'll do all the work for you!
## Introduction video
<br />
```bash
# A quick demonstration of writing an essay with Fabric
```
<video src="https://github.com/danielmiessler/fabric/assets/50654/09c11764-e6ba-4709-952d-450d70d76ac9" controls>
Your browser does not support the video tag.
</video>
<div align="center">
<a href="https://youtu.be/wPEyyigh10g">
<img width="972" alt="fabric_intro_video" src="https://github.com/danielmiessler/fabric/assets/50654/1eb1b9be-0bab-4c77-8ed2-ed265e8a3435">
</a>
</div>
## What and why
@@ -96,7 +93,7 @@ Fabric has Patterns for all sorts of life and work activities, including:
- Getting summaries of long, boring content
- Explaining code to you
- Turning bad documentation into usable documentation
- Create social media posts from any content input
- Creating social media posts from any content input
- And a million more…
### Our approach to prompting
@@ -146,7 +143,13 @@ git clone https://github.com/danielmiessler/fabric.git
cd fabric
```
4. Install poetry
4. Ensure the `setup.sh` script is executable. If you're not sure, you can make it executable by running the following command:
```bash
chmod +x setup.sh
```
5. Install poetry
ref.: https://python-poetry.org/docs/#installing-with-the-official-installer
@@ -154,17 +157,18 @@ ref.: https://python-poetry.org/docs/#installing-with-the-official-installer
curl -sSL https://install.python-poetry.org | python3 -
```
5. Run the `setup.sh`, which will do the following:
- Installs python dependencies.
- Creates aliases in your OS. It should update `~/.bashrc`, `/.zshrc`, and `~/.bash_profile` if they are present in your file system.
6. Run the `setup.sh`, which will do the following:
- Installs python dependencies.
- Creates aliases in your OS. It should update `~/.bashrc`, `/.zshrc`, and `~/.bash_profile` if they are present in your file system.
```bash
./setup.sh
```
6. Restart your shell to reload everything.
7. Restart your shell to reload everything.
7. Set your `OPENAI_API_KEY`.
8. Set your `OPENAI_API_KEY`.
```bash
fabric --setup
@@ -172,7 +176,7 @@ fabric --setup
You'll be asked to enter your OpenAI API key, which will be written to `~/.config/fabric/.env`. Patterns will then be downloaded from Github, which will take a few moments.
8. Now you are up and running! You can test by pulling the help.
9. Now you are up and running! You can test by pulling the help.
```bash
# Making sure the paths are set up correctly
@@ -182,7 +186,6 @@ fabric --help
> [!NOTE]
> If you're using the `server` functions, `fabric-api` and `fabric-webui` need to be run in distinct terminal windows.
### Using the `fabric` client
Once you have it all set up, here's how to use it.
@@ -228,6 +231,12 @@ pbpaste | fabric --pattern summarize
pbpaste | fabric --stream --pattern analyze_claims
```
3. **new** All of the patterns have been added as aliases to your bash (or zsh) config file
```bash
pbpaste | analyze_claims --stream
```
> [!NOTE]
> More examples coming in the next few days, including a demo video!
@@ -253,7 +262,7 @@ The wisdom of crowds for the win.
But we go beyond just providing Patterns. We provide code for you to build your very own Fabric server and personal AI infrastructure!
To get started, head over to the [`/server/`](https://github.com/danielmiessler/fabric/tree/main/server) directory and set up your own Fabric Mill with your own Patterns running! You can then use the [`/client/standalone_client_examples`](https://github.com/danielmiessler/fabric/tree/main/client/standalone_client_examples) to connect to it.
To get started, just run the `./setup.sh` file and it'll set up the client, the API server, and the API server web interface. The output of the setup command will also tell you how to run the commands to start them.
## Structure

View File

@@ -32,7 +32,8 @@ Options include:
--pattern, -p: Select the module for analysis.
--stream, -s: Stream output to another application.
--output, -o: Save the response to a file.
--copy, -c: Copy the response to the clipboard.
--copy, -C: Copy the response to the clipboard.
--context, -c: Use Context file (context.md) to add context to your pattern
Example:

View File

@@ -0,0 +1 @@
3.10

View File

@@ -0,0 +1,81 @@
from langchain_community.tools import DuckDuckGoSearchRun
import os
from crewai import Agent, Task, Crew, Process
from dotenv import load_dotenv
import os
current_directory = os.path.dirname(os.path.realpath(__file__))
config_directory = os.path.expanduser("~/.config/fabric")
env_file = os.path.join(config_directory, ".env")
load_dotenv(env_file)
os.environ['OPENAI_MODEL_NAME'] = 'gpt-4-0125-preview'
# You can choose to use a local model through Ollama for example. See https://docs.crewai.com/how-to/LLM-Connections/ for more information.
# osOPENAI_API_BASE='http://localhost:11434/v1'
# OPENAI_MODEL_NAME='openhermes' # Adjust based on available model
# OPENAI_API_KEY=''
# Install duckduckgo-search for this example:
# !pip install -U duckduckgo-search
search_tool = DuckDuckGoSearchRun()
# Define your agents with roles and goals
researcher = Agent(
role='Senior Research Analyst',
goal='Uncover cutting-edge developments in AI and data science',
backstory="""You work at a leading tech think tank.
Your expertise lies in identifying emerging trends.
You have a knack for dissecting complex data and presenting actionable insights.""",
verbose=True,
allow_delegation=False,
tools=[search_tool]
# You can pass an optional llm attribute specifying what mode you wanna use.
# It can be a local model through Ollama / LM Studio or a remote
# model like OpenAI, Mistral, Antrophic or others (https://docs.crewai.com/how-to/LLM-Connections/)
#
# import os
#
# OR
#
# from langchain_openai import ChatOpenAI
# llm=ChatOpenAI(model_name="gpt-3.5", temperature=0.7)
)
writer = Agent(
role='Tech Content Strategist',
goal='Craft compelling content on tech advancements',
backstory="""You are a renowned Content Strategist, known for your insightful and engaging articles.
You transform complex concepts into compelling narratives.""",
verbose=True,
allow_delegation=True
)
# Create tasks for your agents
task1 = Task(
description="""Conduct a comprehensive analysis of the latest advancements in AI in 2024.
Identify key trends, breakthrough technologies, and potential industry impacts.""",
expected_output="Full analysis report in bullet points",
agent=researcher
)
task2 = Task(
description="""Using the insights provided, develop an engaging blog
post that highlights the most significant AI advancements.
Your post should be informative yet accessible, catering to a tech-savvy audience.
Make it sound cool, avoid complex words so it doesn't sound like AI.""",
expected_output="Full blog post of at least 4 paragraphs",
agent=writer
)
# Instantiate your crew with a sequential process
crew = Crew(
agents=[researcher, writer],
tasks=[task1, task2],
verbose=2, # You can set it to 1 or 2 to different logging levels
)
# Get your crew to work!
result = crew.kickoff()
print("######################")
print(result)

View File

@@ -0,0 +1,89 @@
from crewai import Crew
from textwrap import dedent
from .trip_agents import TripAgents
from .trip_tasks import TripTasks
import os
from dotenv import load_dotenv
current_directory = os.path.dirname(os.path.realpath(__file__))
config_directory = os.path.expanduser("~/.config/fabric")
env_file = os.path.join(config_directory, ".env")
load_dotenv(env_file)
os.environ['OPENAI_MODEL_NAME'] = 'gpt-4-0125-preview'
class TripCrew:
def __init__(self, origin, cities, date_range, interests):
self.cities = cities
self.origin = origin
self.interests = interests
self.date_range = date_range
def run(self):
agents = TripAgents()
tasks = TripTasks()
city_selector_agent = agents.city_selection_agent()
local_expert_agent = agents.local_expert()
travel_concierge_agent = agents.travel_concierge()
identify_task = tasks.identify_task(
city_selector_agent,
self.origin,
self.cities,
self.interests,
self.date_range
)
gather_task = tasks.gather_task(
local_expert_agent,
self.origin,
self.interests,
self.date_range
)
plan_task = tasks.plan_task(
travel_concierge_agent,
self.origin,
self.interests,
self.date_range
)
crew = Crew(
agents=[
city_selector_agent, local_expert_agent, travel_concierge_agent
],
tasks=[identify_task, gather_task, plan_task],
verbose=True
)
result = crew.kickoff()
return result
class planner_cli:
def ask(self):
print("## Welcome to Trip Planner Crew")
print('-------------------------------')
location = input(
dedent("""
From where will you be traveling from?
"""))
cities = input(
dedent("""
What are the cities options you are interested in visiting?
"""))
date_range = input(
dedent("""
What is the date range you are interested in traveling?
"""))
interests = input(
dedent("""
What are some of your high level interests and hobbies?
"""))
trip_crew = TripCrew(location, cities, date_range, interests)
result = trip_crew.run()
print("\n\n########################")
print("## Here is you Trip Plan")
print("########################\n")
print(result)

View File

@@ -0,0 +1,38 @@
import json
import os
import requests
from crewai import Agent, Task
from langchain.tools import tool
from unstructured.partition.html import partition_html
class BrowserTools():
@tool("Scrape website content")
def scrape_and_summarize_website(website):
"""Useful to scrape and summarize a website content"""
url = f"https://chrome.browserless.io/content?token={os.environ['BROWSERLESS_API_KEY']}"
payload = json.dumps({"url": website})
headers = {'cache-control': 'no-cache', 'content-type': 'application/json'}
response = requests.request("POST", url, headers=headers, data=payload)
elements = partition_html(text=response.text)
content = "\n\n".join([str(el) for el in elements])
content = [content[i:i + 8000] for i in range(0, len(content), 8000)]
summaries = []
for chunk in content:
agent = Agent(
role='Principal Researcher',
goal=
'Do amazing researches and summaries based on the content you are working with',
backstory=
"You're a Principal Researcher at a big company and you need to do a research about a given topic.",
allow_delegation=False)
task = Task(
agent=agent,
description=
f'Analyze and summarize the content bellow, make sure to include the most relevant information in the summary, return only the summary nothing else.\n\nCONTENT\n----------\n{chunk}'
)
summary = task.execute()
summaries.append(summary)
return "\n\n".join(summaries)

View File

@@ -0,0 +1,15 @@
from langchain.tools import tool
class CalculatorTools():
@tool("Make a calculation")
def calculate(operation):
"""Useful to perform any mathematical calculations,
like sum, minus, multiplication, division, etc.
The input to this tool should be a mathematical
expression, a couple examples are `200*7` or `5000/2*10`
"""
try:
return eval(operation)
except SyntaxError:
return "Error: Invalid syntax in mathematical expression"

View File

@@ -0,0 +1,37 @@
import json
import os
import requests
from langchain.tools import tool
class SearchTools():
@tool("Search the internet")
def search_internet(query):
"""Useful to search the internet
about a a given topic and return relevant results"""
top_result_to_return = 4
url = "https://google.serper.dev/search"
payload = json.dumps({"q": query})
headers = {
'X-API-KEY': os.environ['SERPER_API_KEY'],
'content-type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
# check if there is an organic key
if 'organic' not in response.json():
return "Sorry, I couldn't find anything about that, there could be an error with you serper api key."
else:
results = response.json()['organic']
string = []
for result in results[:top_result_to_return]:
try:
string.append('\n'.join([
f"Title: {result['title']}", f"Link: {result['link']}",
f"Snippet: {result['snippet']}", "\n-----------------"
]))
except KeyError:
next
return '\n'.join(string)

View File

@@ -0,0 +1,45 @@
from crewai import Agent
from .tools.browser_tools import BrowserTools
from .tools.calculator_tools import CalculatorTools
from .tools.search_tools import SearchTools
class TripAgents():
def city_selection_agent(self):
return Agent(
role='City Selection Expert',
goal='Select the best city based on weather, season, and prices',
backstory='An expert in analyzing travel data to pick ideal destinations',
tools=[
SearchTools.search_internet,
BrowserTools.scrape_and_summarize_website,
],
verbose=True)
def local_expert(self):
return Agent(
role='Local Expert at this city',
goal='Provide the BEST insights about the selected city',
backstory="""A knowledgeable local guide with extensive information
about the city, it's attractions and customs""",
tools=[
SearchTools.search_internet,
BrowserTools.scrape_and_summarize_website,
],
verbose=True)
def travel_concierge(self):
return Agent(
role='Amazing Travel Concierge',
goal="""Create the most amazing travel itineraries with budget and
packing suggestions for the city""",
backstory="""Specialist in travel planning and logistics with
decades of experience""",
tools=[
SearchTools.search_internet,
BrowserTools.scrape_and_summarize_website,
CalculatorTools.calculate,
],
verbose=True)

View File

@@ -0,0 +1,83 @@
from crewai import Task
from textwrap import dedent
from datetime import date
class TripTasks():
def identify_task(self, agent, origin, cities, interests, range):
return Task(description=dedent(f"""
Analyze and select the best city for the trip based
on specific criteria such as weather patterns, seasonal
events, and travel costs. This task involves comparing
multiple cities, considering factors like current weather
conditions, upcoming cultural or seasonal events, and
overall travel expenses.
Your final answer must be a detailed
report on the chosen city, and everything you found out
about it, including the actual flight costs, weather
forecast and attractions.
{self.__tip_section()}
Traveling from: {origin}
City Options: {cities}
Trip Date: {range}
Traveler Interests: {interests}
"""),
agent=agent)
def gather_task(self, agent, origin, interests, range):
return Task(description=dedent(f"""
As a local expert on this city you must compile an
in-depth guide for someone traveling there and wanting
to have THE BEST trip ever!
Gather information about key attractions, local customs,
special events, and daily activity recommendations.
Find the best spots to go to, the kind of place only a
local would know.
This guide should provide a thorough overview of what
the city has to offer, including hidden gems, cultural
hotspots, must-visit landmarks, weather forecasts, and
high level costs.
The final answer must be a comprehensive city guide,
rich in cultural insights and practical tips,
tailored to enhance the travel experience.
{self.__tip_section()}
Trip Date: {range}
Traveling from: {origin}
Traveler Interests: {interests}
"""),
agent=agent)
def plan_task(self, agent, origin, interests, range):
return Task(description=dedent(f"""
Expand this guide into a a full 7-day travel
itinerary with detailed per-day plans, including
weather forecasts, places to eat, packing suggestions,
and a budget breakdown.
You MUST suggest actual places to visit, actual hotels
to stay and actual restaurants to go to.
This itinerary should cover all aspects of the trip,
from arrival to departure, integrating the city guide
information with practical travel logistics.
Your final answer MUST be a complete expanded travel plan,
formatted as markdown, encompassing a daily schedule,
anticipated weather conditions, recommended clothing and
items to pack, and a detailed budget, ensuring THE BEST
TRIP EVER, Be specific and give it a reason why you picked
# up each place, what make them special! {self.__tip_section()}
Trip Date: {range}
Traveling from: {origin}
Traveler Interests: {interests}
"""),
agent=agent)
def __tip_section(self):
return "If you do your BEST WORK, I'll tip you $100!"

View File

@@ -0,0 +1,3 @@
# Context
please give all responses in spanish

View File

@@ -1,19 +1,27 @@
from .utils import Standalone, Update, Setup
from .utils import Standalone, Update, Setup, Alias, AgentSetup
import argparse
import sys
import time
import os
script_directory = os.path.dirname(os.path.realpath(__file__))
def main():
parser = argparse.ArgumentParser(
description="An open source framework for augmenting humans using AI."
)
parser.add_argument("--text", "-t", help="Text to extract summary from")
parser.add_argument(
"--copy", "-c", help="Copy the response to the clipboard", action="store_true"
"--copy", "-C", help="Copy the response to the clipboard", action="store_true"
)
subparsers = parser.add_subparsers(dest='command', help='Sub-command help')
agents_parser = subparsers.add_parser('agents', help='Crew command help')
agents_parser.add_argument(
"trip_planner", help="The origin city for the trip")
agents_parser.add_argument(
'ApiKeys', help="enter API keys for tools", action="store_true")
parser.add_argument(
"--output",
"-o",
@@ -31,7 +39,8 @@ def main():
parser.add_argument(
"--list", "-l", help="List available patterns", action="store_true"
)
parser.add_argument("--update", "-u", help="Update patterns", action="store_true")
parser.add_argument(
"--update", "-u", help="Update patterns", action="store_true")
parser.add_argument("--pattern", "-p", help="The pattern (prompt) to use")
parser.add_argument(
"--setup", help="Set up your fabric instance", action="store_true"
@@ -42,31 +51,54 @@ def main():
parser.add_argument(
"--listmodels", help="List all available models", action="store_true"
)
parser.add_argument('--context', '-c',
help="Use Context file (context.md) to add context to your pattern", action="store_true")
args = parser.parse_args()
home_holder = os.path.expanduser("~")
config = os.path.join(home_holder, ".config", "fabric")
config_patterns_directory = os.path.join(config, "patterns")
config_context = os.path.join(config, "context.md")
env_file = os.path.join(config, ".env")
if not os.path.exists(config):
os.makedirs(config)
if args.setup:
Setup().run()
Alias()
sys.exit()
if not os.path.exists(env_file) or not os.path.exists(config_patterns_directory):
print("Please run --setup to set up your API key and download patterns.")
sys.exit()
if not os.path.exists(config_patterns_directory):
Update()
Alias()
sys.exit()
if args.command == "agents":
from .agents.trip_planner.main import planner_cli
if args.ApiKeys:
AgentSetup().apiKeys()
sys.exit()
if not args.trip_planner:
print("Please provide an agent")
print(f"Available Agents:")
for agent in tripcrew.agents:
print(agent)
else:
tripcrew = planner_cli()
tripcrew.ask()
sys.exit()
if args.update:
Update()
print("Your Patterns have been updated.")
Alias()
sys.exit()
if args.context:
if not os.path.exists(os.path.join(config, "context.md")):
print("Please create a context.md file in ~/.config/fabric")
sys.exit()
standalone = Standalone(args, args.pattern)
if args.list:
try:
direct = os.listdir(config_patterns_directory)
direct = sorted(os.listdir(config_patterns_directory))
for d in direct:
print(d)
sys.exit()
@@ -80,10 +112,23 @@ def main():
text = args.text
else:
text = standalone.get_cli_input()
if args.stream:
if args.stream and not args.context:
standalone.streamMessage(text)
sys.exit()
if args.stream and args.context:
with open(config_context, "r") as f:
context = f.read()
standalone.streamMessage(text, context=context)
sys.exit()
elif args.context:
with open(config_context, "r") as f:
context = f.read()
standalone.sendMessage(text, context=context)
sys.exit()
else:
standalone.sendMessage(text)
sys.exit()
if __name__ == "__main__":
main()

View File

@@ -7,13 +7,15 @@ import platform
from dotenv import load_dotenv
from requests.exceptions import HTTPError
from tqdm import tqdm
import zipfile
import tempfile
import shutil
current_directory = os.path.dirname(os.path.realpath(__file__))
config_directory = os.path.expanduser("~/.config/fabric")
env_file = os.path.join(config_directory, ".env")
class Standalone:
def __init__(self, args, pattern="", env_file="~/.config/fabric/.env"):
""" Initialize the class with the provided arguments and environment file.
@@ -49,7 +51,7 @@ class Standalone:
self.args = args
self.model = args.model
def streamMessage(self, input_data: str):
def streamMessage(self, input_data: str, context=""):
""" Stream a message and handle exceptions.
Args:
@@ -71,14 +73,21 @@ class Standalone:
if self.pattern:
try:
with open(wisdom_File, "r") as f:
system = f.read()
if context:
system = context + '\n\n' + f.read()
else:
system = f.read()
system_message = {"role": "system", "content": system}
messages = [system_message, user_message]
except FileNotFoundError:
print("pattern not found")
return
else:
messages = [user_message]
if context:
messages = [
{"role": "system", "content": context}, user_message]
else:
messages = [user_message]
try:
stream = self.client.chat.completions.create(
model=self.model,
@@ -109,7 +118,7 @@ class Standalone:
with open(self.args.output, "w") as f:
f.write(buffer)
def sendMessage(self, input_data: str):
def sendMessage(self, input_data: str, context=""):
""" Send a message using the input data and generate a response.
Args:
@@ -130,14 +139,21 @@ class Standalone:
if self.pattern:
try:
with open(wisdom_File, "r") as f:
system = f.read()
if context:
system = context + '\n\n' + f.read()
else:
system = f.read()
system_message = {"role": "system", "content": system}
messages = [system_message, user_message]
except FileNotFoundError:
print("pattern not found")
return
else:
messages = [user_message]
if context:
messages = [
{'role': 'system', 'content': context}, user_message]
else:
messages = [user_message]
try:
response = self.client.chat.completions.create(
model=self.model,
@@ -159,28 +175,30 @@ class Standalone:
def fetch_available_models(self):
headers = {
"Authorization": f"Bearer { self.client.api_key }"
"Authorization": f"Bearer {self.client.api_key}"
}
response = requests.get("https://api.openai.com/v1/models", headers=headers)
response = requests.get(
"https://api.openai.com/v1/models", headers=headers)
if response.status_code == 200:
models = response.json().get("data", [])
# Filter only gpt models
gpt_models = [model for model in models if model.get("id", "").startswith(("gpt"))]
gpt_models = [model for model in models if model.get(
"id", "").startswith(("gpt"))]
# Sort the models alphabetically by their ID
sorted_gpt_models = sorted(gpt_models, key=lambda x: x.get("id"))
for model in sorted_gpt_models:
print(model.get("id"))
else:
print(f"Failed to fetch models: HTTP {response.status_code}")
def get_cli_input(self):
""" aided by ChatGPT; uses platform library
accepts either piped input or console input
from either Windows or Linux
Args:
none
Returns:
@@ -191,129 +209,112 @@ class Standalone:
if not sys.stdin.isatty(): # Check if input is being piped
return sys.stdin.read().strip() # Read piped input
else:
return input("Enter Question: ") # Prompt user for input from console
# Prompt user for input from console
return input("Enter Question: ")
else:
return sys.stdin.read()
class Update:
def __init__(self):
""" Initialize the object with default values and update patterns.
This method initializes the object with default values for root_api_url, config_directory, and pattern_directory.
It then creates the pattern_directory if it does not exist and calls the update_patterns method to update the patterns.
Raises:
OSError: If there is an issue creating the pattern_directory.
"""
self.root_api_url = "https://api.github.com/repos/danielmiessler/fabric/contents/patterns?ref=main"
"""Initialize the object with default values."""
self.repo_zip_url = "https://github.com/danielmiessler/fabric/archive/refs/heads/main.zip"
self.config_directory = os.path.expanduser("~/.config/fabric")
self.pattern_directory = os.path.join(self.config_directory, "patterns")
self.pattern_directory = os.path.join(
self.config_directory, "patterns")
os.makedirs(self.pattern_directory, exist_ok=True)
self.update_patterns() # Call the update process from a method.
print("Updating patterns...")
self.update_patterns() # Start the update process immediately
def update_patterns(self):
""" Update the patterns by downloading from the GitHub directory.
Raises:
HTTPError: If there is an HTTP error while downloading patterns.
"""
try:
self.progress_bar = tqdm(desc="Downloading Patterns…", unit="file")
self.get_github_directory_contents(
self.root_api_url, self.pattern_directory
)
# Close progress bar on success before printing the message.
self.progress_bar.close()
except HTTPError as e:
# Ensure progress bar is closed on HTTPError as well.
self.progress_bar.close()
if e.response.status_code == 403:
print(
"GitHub API rate limit exceeded. Please wait before trying again."
)
sys.exit()
"""Update the patterns by downloading the zip from GitHub and extracting it."""
with tempfile.TemporaryDirectory() as temp_dir:
zip_path = os.path.join(temp_dir, "repo.zip")
self.download_zip(self.repo_zip_url, zip_path)
extracted_folder_path = self.extract_zip(zip_path, temp_dir)
# The patterns folder will be inside "fabric-main" after extraction
patterns_source_path = os.path.join(
extracted_folder_path, "fabric-main", "patterns")
if os.path.exists(patterns_source_path):
# If the patterns directory already exists, remove it before copying over the new one
if os.path.exists(self.pattern_directory):
shutil.rmtree(self.pattern_directory)
shutil.copytree(patterns_source_path, self.pattern_directory)
print("Patterns updated successfully.")
else:
print(f"Failed to download patterns due to an HTTP error: {e}")
sys.exit() # Exit after handling the error.
print("Patterns folder not found in the downloaded zip.")
def download_file(self, url, local_path):
""" Download a file from the given URL and save it to the local path.
def download_zip(self, url, save_path):
"""Download the zip file from the specified URL."""
response = requests.get(url)
response.raise_for_status() # Check if the download was successful
with open(save_path, 'wb') as f:
f.write(response.content)
print("Downloaded zip file successfully.")
Args:
url (str): The URL of the file to be downloaded.
local_path (str): The local path where the file will be saved.
def extract_zip(self, zip_path, extract_to):
"""Extract the zip file to the specified directory."""
with zipfile.ZipFile(zip_path, 'r') as zip_ref:
zip_ref.extractall(extract_to)
print("Extracted zip file successfully.")
return extract_to # Return the path to the extracted contents
Raises:
HTTPError: If an HTTP error occurs during the download process.
"""
try:
response = requests.get(url)
response.raise_for_status()
with open(local_path, "wb") as f:
f.write(response.content)
self.progress_bar.update(1)
except HTTPError as e:
print(f"Failed to download file {url}. HTTP error: {e}")
sys.exit()
class Alias:
def __init__(self):
self.config_files = []
home_directory = os.path.expanduser("~")
self.patterns = os.path.join(home_directory, ".config/fabric/patterns")
if os.path.exists(os.path.join(home_directory, ".bashrc")):
self.config_files.append(os.path.join(home_directory, ".bashrc"))
if os.path.exists(os.path.join(home_directory, ".zshrc")):
self.config_files.append(os.path.join(home_directory, ".zshrc"))
if os.path.exists(os.path.join(home_directory, ".bash_profile")):
self.config_files.append(os.path.join(
home_directory, ".bash_profile"))
self.remove_all_patterns()
self.add_patterns()
print('Aliases added successfully. Please restart your terminal to use them.')
def process_item(self, item, local_dir):
""" Process the given item and save it to the local directory.
def add(self, name, alias):
for file in self.config_files:
with open(file, "a") as f:
f.write(f"alias {name}='{alias}'\n")
Args:
item (dict): The item to be processed, containing information about the type, download URL, name, and URL.
local_dir (str): The local directory where the item will be saved.
def remove(self, pattern):
for file in self.config_files:
# Read the whole file first
with open(file, "r") as f:
wholeFile = f.read()
Returns:
None
# Determine if the line to be removed is in the file
target_line = f"alias {pattern}='fabric --pattern {pattern}'\n"
if target_line in wholeFile:
# If the line exists, replace it with nothing (remove it)
wholeFile = wholeFile.replace(target_line, "")
Raises:
OSError: If there is an issue creating the new directory using os.makedirs.
"""
# Write the modified content back to the file
with open(file, "w") as f:
f.write(wholeFile)
if item["type"] == "file":
self.download_file(
item["download_url"], os.path.join(local_dir, item["name"])
)
elif item["type"] == "dir":
new_dir = os.path.join(local_dir, item["name"])
os.makedirs(new_dir, exist_ok=True)
self.get_github_directory_contents(item["url"], new_dir)
def remove_all_patterns(self):
allPatterns = os.listdir(self.patterns)
for pattern in allPatterns:
self.remove(pattern)
def get_github_directory_contents(self, api_url, local_dir):
""" Get the contents of a directory from GitHub API and process each item.
def find_line(self, name):
for file in self.config_files:
with open(file, "r") as f:
lines = f.readlines()
for line in lines:
if line.strip("\n") == f"alias ${name}='{alias}'":
return line
Args:
api_url (str): The URL of the GitHub API endpoint for the directory.
local_dir (str): The local directory where the contents will be processed.
def add_patterns(self):
allPatterns = os.listdir(self.patterns)
for pattern in allPatterns:
self.add(pattern, f"fabric --pattern {pattern}")
Returns:
None
Raises:
HTTPError: If an HTTP error occurs while fetching the directory contents.
If the status code is 403, it prints a message about GitHub API rate limit exceeded
and closes the progress bar. For any other status code, it prints a message
about failing to fetch directory contents due to an HTTP error.
"""
try:
response = requests.get(api_url)
response.raise_for_status()
jsonList = response.json()
for item in jsonList:
self.process_item(item, local_dir)
except HTTPError as e:
if e.response.status_code == 403:
print(
"GitHub API rate limit exceeded. Please wait before trying again."
)
self.progress_bar.close() # Ensure the progress bar is cleaned up properly
else:
print(f"Failed to fetch directory contents due to an HTTP error: {e}")
class Setup:
def __init__(self):
@@ -324,7 +325,8 @@ class Setup:
"""
self.config_directory = os.path.expanduser("~/.config/fabric")
self.pattern_directory = os.path.join(self.config_directory, "patterns")
self.pattern_directory = os.path.join(
self.config_directory, "patterns")
os.makedirs(self.pattern_directory, exist_ok=True)
self.env_file = os.path.join(self.config_directory, ".env")
@@ -354,7 +356,6 @@ class Setup:
"""
Update()
sys.exit()
def run(self):
""" Execute the Fabric program.
@@ -370,25 +371,25 @@ class Setup:
self.api_key(apikey.strip())
self.patterns()
class Transcribe:
def youtube(video_id):
"""
This method gets the transciption
of a YouTube video designated with the video_id
Input:
the video id specifing a YouTube video
an example url for a video: https://www.youtube.com/watch?v=vF-MQmVxnCs&t=306s
the video id is vF-MQmVxnCs&t=306s
Output:
a transcript for the video
Raises:
an exception and prints error
"""
try:
transcript_list = YouTubeTranscriptApi.get_transcript(video_id)
@@ -399,5 +400,32 @@ class Transcribe:
except Exception as e:
print("Error:", e)
return None
class AgentSetup:
def apiKeys(self):
"""Method to set the API keys in the environment file.
Returns:
None
"""
print("Welcome to Fabric. Let's get started.")
browserless = input("Please enter your Browserless API key\n")
serper = input("Please enter your Serper API key\n")
# Entries to be added
browserless_entry = f"BROWSERLESS_API_KEY={browserless}"
serper_entry = f"SERPER_API_KEY={serper}"
# Check and write to the file
with open(env_file, "r+") as f:
content = f.read()
# Determine if the file ends with a newline
if content.endswith('\n'):
# If it ends with a newline, we directly write the new entries
f.write(f"{browserless_entry}\n{serper_entry}\n")
else:
# If it does not end with a newline, add one before the new entries
f.write(f"\n{browserless_entry}\n{serper_entry}\n")

BIN
patterns/.DS_Store vendored

Binary file not shown.

View File

@@ -6,58 +6,37 @@ Take a deep breath and think step by step about how to best accomplish this goal
# OUTPUT SECTIONS
- Extract a summary of the content in 50 words or less, including who is presenting and the content being discussed into a section called SUMMARY.
- Extract a summary of the paper and its conclusions in into a 25-word sentence called SUMMARY.
- Extract the list of authors in a section called AUTHORS.
- Extract the list of organizations the authors are associated, e.g., which university they're at, with in a section called AUTHOR ORGANIZATIONS.
- Extract the primary paper findings into a bulleted list of no more than 50 words per bullet into a section called FINDINGS.
- Extract the primary paper findings into a bulleted list of no more than 25 words per bullet into a section called FINDINGS.
- You extract the size and details of the study for the research in a section called STUDY DETAILS.
- Extract the overall structure and character of the study for the research in a section called STUDY DETAILS.
- Extract the study quality by evaluating the following items in a section called STUDY QUALITY:
- Extract the study quality by evaluating the following items in a section called STUDY QUALITY that has the following sub-sections:
### Sample size
- Study Design: (give a 25 word description, including the pertinent data and statistics.)
- Sample Size: (give a 25 word description, including the pertinent data and statistics.)
- Confidence Intervals (give a 25 word description, including the pertinent data and statistics.)
- P-value (give a 25 word description, including the pertinent data and statistics.)
- Effect Size (give a 25 word description, including the pertinent data and statistics.)
- Consistency of Results (give a 25 word description, including the pertinent data and statistics.)
- Data Analysis Method (give a 25 word description, including the pertinent data and statistics.)
- **Check the Sample Size**: The larger the sample size, the more confident you can be in the findings. A larger sample size reduces the margin of error and increases the study's power.
- Discuss any Conflicts of Interest in a section called CONFLICTS OF INTEREST. Rate the conflicts of interest as NONE DETECTED, LOW, MEDIUM, HIGH, or CRITICAL.
### Confidence intervals
- Extract the researcher's analysis and interpretation in a section called RESEARCHER'S INTERPRETATION, including how confident they are in the results being real and likely to be replicated on a scale of LOW, MEDIUM, or HIGH.
- **Look at the Confidence Intervals**: Confidence intervals provide a range within which the true population parameter lies with a certain degree of confidence (usually 95% or 99%). Narrower confidence intervals suggest a higher level of precision and confidence in the estimate.
### P-Value
- **Evaluate the P-value**: The P-value tells you the probability that the results occurred by chance. A lower P-value (typically less than 0.05) suggests that the findings are statistically significant and not due to random chance.
### Effect size
- **Consider the Effect Size**: Effect size tells you how much of a difference there is between groups. A larger effect size indicates a stronger relationship and more confidence in the findings.
### Study design
- **Review the Study Design**: Randomized controlled trials are usually considered the gold standard in research. If the study is observational, it may be less reliable.
### Consistency of results
- **Check for Consistency of Results**: If the results are consistent across multiple studies, it increases the confidence in the findings.
### Data analysis methods
- **Examine the Data Analysis Methods**: Check if the data analysis methods used are appropriate for the type of data and research question. Misuse of statistical methods can lead to incorrect conclusions.
### Researcher's interpretation
- **Assess the Researcher's Interpretation**: The researchers should interpret their results in the context of the study's limitations. Overstating the findings can misrepresent the confidence level.
### Summary
You output a 50 word summary of the quality of the paper and it's likelihood of being replicated in future work as one of three levels: High, Medium, or Low. You put that sentence and ratign into a section called SUMMARY.
- Based on all of the analysis performed above, output a 25 word summary of the quality of the paper and it's likelihood of being replicated in future work as one of five levels: VERY LOW, LOW, MEDIUM, HIGH, or VERY HIGH. You put that sentence and RATING into a section called SUMMARY and RATING.
# OUTPUT INSTRUCTIONS
- Create the output using the formatting above.
- You only output human readable Markdown.
- In the markdown, don't use formatting like bold or italics. Make the output maximially readable in plain text.
- Do not output warnings or notes—just the requested sections.
# INPUT:

View File

@@ -0,0 +1,38 @@
# IDENTITY and PURPOSE
You are a super-intelligent cybersecurity expert. You specialize in extracting the surprising, insightful, and interesting information from cybersecurity threat reports.
Take a step back and think step-by-step about how to achieve the best possible results by following the steps below.
# STEPS
- Read the entire threat report from an expert perspective, thinking deeply about what's new, interesting, and surprising in the report.
- Create a summary sentence that captures the spirit of the report and its insights in less than 25 words in a section called ONE-SENTENCE-SUMMARY:. Use plain and conversational language when creating this summary. Don't use jargon or marketing language.
- Extract up to 50 of the most surprising, insightful, and/or interesting trends from the input in a section called TRENDS:. If there are less than 50 then collect all of them. Make sure you extract at least 20.
- Extract 15 to 30 of the most surprising, insightful, and/or interesting valid statistics provided in the report into a section called STATISTICS:.
- Extract 15 to 30 of the most surprising, insightful, and/or interesting quotes from the input into a section called QUOTES:. Use the exact quote text from the input.
- Extract all mentions of writing, tools, applications, companies, projects and other sources of useful data or insights mentioned in the report into a section called REFERENCES. This should include any and all references to something that the report mentioned.
- Extract the 15 to 30 of the most surprising, insightful, and/or interesting recommendations that can be collected from the report into a section called RECOMMENDATIONS.
# OUTPUT INSTRUCTIONS
- Only output Markdown.
- Do not output the markdown code syntax, only the content.
- Do not use bold or italics formatting in the markdown output.
- Extract at least 20 TRENDS from the content.
- Extract at least 10 items for the other output sections.
- Do not give warnings or notes; only output the requested sections.
- You use bulleted lists for output, not numbered lists.
- Do not repeat ideas, quotes, facts, or resources.
- Do not start items with the same opening words.
- Ensure you follow ALL these instructions when creating your output.
# INPUT
INPUT:

View File

@@ -0,0 +1 @@
CONTENT:

View File

@@ -0,0 +1,27 @@
# IDENTITY and PURPOSE
You are a super-intelligent cybersecurity expert. You specialize in extracting the surprising, insightful, and interesting information from cybersecurity threat reports.
Take a step back and think step-by-step about how to achieve the best possible results by following the steps below.
# STEPS
- Read the entire threat report from an expert perspective, thinking deeply about what's new, interesting, and surprising in the report.
- Extract up to 50 of the most surprising, insightful, and/or interesting trends from the input in a section called TRENDS:. If there are less than 50 then collect all of them. Make sure you extract at least 20.
# OUTPUT INSTRUCTIONS
- Only output Markdown.
- Do not output the markdown code syntax, only the content.
- Do not use bold or italics formatting in the markdown output.
- Extract at least 20 TRENDS from the content.
- Do not give warnings or notes; only output the requested sections.
- You use bulleted lists for output, not numbered lists.
- Do not repeat ideas, quotes, facts, or resources.
- Do not start items with the same opening words.
- Ensure you follow ALL these instructions when creating your output.
# INPUT
INPUT:

View File

@@ -0,0 +1 @@
CONTENT:

View File

@@ -0,0 +1,62 @@
# IDENTITY and PURPOSE
You are an expert conversation topic and timestamp creator. You take a transcript and you extract the most interesting topics discussed and give timestamps for where in the video they occur.
Take a step back and think step-by-step about how you would do this. You would probably start by "watching" the video (via the transcript) and taking notes on the topics discussed and the time they were discussed. Then you would take those notes and create a list of topics and timestamps.
# STEPS
- Fully consume the transcript as if you're watching or listening to the content.
- Think deeply about the topics discussed and what were the most interesting subjects and moments in the content.
- Name those subjects and/moments in 2-3 capitalized words.
- Match the timestamps to the topics. Note that input timestamps have the following format: HOURS:MINUTES:SECONDS.MILLISECONDS, which is not the same as the OUTPUT format!
INPUT SAMPLE
[02:17:43.120 --> 02:17:49.200] same way. I'll just say the same. And I look forward to hearing the response to my job application
[02:17:49.200 --> 02:17:55.040] that I've submitted. Oh, you're accepted. Oh, yeah. We all speak of you all the time. Thank you so
[02:17:55.040 --> 02:18:00.720] much. Thank you, guys. Thank you. Thanks for listening to this conversation with Neri Oxman.
[02:18:00.720 --> 02:18:05.520] To support this podcast, please check out our sponsors in the description. And now,
END INPUT SAMPLE
The OUTPUT TIMESTAMP format is:
00:00:00 (HOURS:MINUTES:SECONDS) (HH:MM:SS)
- Note the maximum length of the video based on the last timestamp.
- Ensure all output timestamps are sequential and fall within the length of the content.
# OUTPUT INSTRUCTIONS
EXAMPLE OUTPUT (Hours:Minutes:Seconds)
00:00:00 Members-only Forum Access
00:00:10 Live Hacking Demo
00:00:26 Ideas vs. Book
00:00:30 Meeting Will Smith
00:00:44 How to Influence Others
00:01:34 Learning by Reading
00:58:30 Writing With Punch
00:59:22 100 Posts or GTFO
01:00:32 How to Gain Followers
01:01:31 The Music That Shapes
01:27:21 Subdomain Enumeration Demo
01:28:40 Hiding in Plain Sight
01:29:06 The Universe Machine
00:09:36 Early School Experiences
00:10:12 The First Business Failure
00:10:32 David Foster Wallace
00:12:07 Copying Other Writers
00:12:32 Practical Advice for N00bs
END EXAMPLE OUTPUT
- Ensure all output timestamps are sequential and fall within the length of the content, e.g., if the total length of the video is 24 minutes. (00:00:00 - 00:24:00), then no output can be 01:01:25, or anything over 00:25:00 or over!
- ENSURE the output timestamps and topics are shown gradually and evenly incrementing from 00:00:00 to the final timestamp of the content.
INPUT:

View File

View File

@@ -0,0 +1,154 @@
<div align="center">
<img src="https://beehiiv-images-production.s3.amazonaws.com/uploads/asset/file/2012aa7c-a939-4262-9647-7ab614e02601/extwis-logo-miessler.png?t=1704502975" alt="extwislogo" width="400" height="400"/>
# `/extractwisdom`
<h4><code>extractwisdom</code> is a <a href="https://github.com/danielmiessler/fabric" target="_blank">Fabric</a> pattern that <em>extracts wisdom</em> from any text.</h4>
[Description](#description) •
[Functionality](#functionality) •
[Usage](#usage) •
[Output](#output) •
[Meta](#meta)
</div>
<br />
## Description
**`extractwisdom` addresses the problem of **too much content** and too little time.**
_Not only that, but it's also too easy to forget the stuff read, watch, or listen to._
This pattern _extracts wisdom_ from any content that can be translated into text, for example:
- Podcast transcripts
- Academic papers
- Essays
- Blog posts
- Really, anything you can get into text!
## Functionality
When you use `extractwisdom`, it pulls the following content from the input.
- `IDEAS`
- Extracts the best ideas from the content, i.e., what you might have taken notes on if you were doing so manually.
- `QUOTES`
- Some of the best quotes from the content.
- `REFERENCES`
- External writing, art, and other content referenced positively during the content that might be worth following up on.
- `HABITS`
- Habits of the speakers that could be worth replicating.
- `RECOMMENDATIONS`
- A list of things that the content recommends Habits of the speakers.
### Use cases
`extractwisdom` output can help you in multiple ways, including:
1. `Time Filtering`<br />
Allows you to quickly see if content is worth an in-depth review or not.
2. `Note Taking`<br />
Can be used as a substitute for taking time-consuming, manual notes on the content.
## Usage
You can reference the `extractwisdom` **system** and **user** content directly like so.
### Pull the _system_ prompt directly
```sh
curl -sS https://github.com/danielmiessler/fabric/blob/main/extract-wisdom/dmiessler/extract-wisdom-1.0.0/system.md
```
### Pull the _user_ prompt directly
```sh
curl -sS https://github.com/danielmiessler/fabric/blob/main/extract-wisdom/dmiessler/extract-wisdom-1.0.0/user.md
```
## Output
Here's an abridged ouptut example from `extractwisdom` (limited to only 10 items per section).
```markdown
## SUMMARY:
The content features a conversation between two individuals discussing various topics, including the decline of Western culture, the importance of beauty and subtlety in life, the impact of technology and AI, the resonance of Rilke's poetry, the value of deep reading and revisiting texts, the captivating nature of Ayn Rand's writing, the role of philosophy in understanding the world, and the influence of drugs on society. They also touch upon creativity, attention spans, and the importance of introspection.
## IDEAS:
1. Western culture is perceived to be declining due to a loss of values and an embrace of mediocrity.
2. Mass media and technology have contributed to shorter attention spans and a need for constant stimulation.
3. Rilke's poetry resonates due to its focus on beauty and ecstasy in everyday objects.
4. Subtlety is often overlooked in modern society due to sensory overload.
5. The role of technology in shaping music and performance art is significant.
6. Reading habits have shifted from deep, repetitive reading to consuming large quantities of new material.
7. Revisiting influential books as one ages can lead to new insights based on accumulated wisdom and experiences.
8. Fiction can vividly illustrate philosophical concepts through characters and narratives.
9. Many influential thinkers have backgrounds in philosophy, highlighting its importance in shaping reasoning skills.
10. Philosophy is seen as a bridge between theology and science, asking questions that both fields seek to answer.
## QUOTES:
1. "You can't necessarily think yourself into the answers. You have to create space for the answers to come to you."
2. "The West is dying and we are killing her."
3. "The American Dream has been replaced by mass packaged mediocrity porn, encouraging us to revel like happy pigs in our own meekness."
4. "There's just not that many people who have the courage to reach beyond consensus and go explore new ideas."
5. "I'll start watching Netflix when I've read the whole of human history."
6. "Rilke saw beauty in everything... He sees it's in one little thing, a representation of all things that are beautiful."
7. "Vanilla is a very subtle flavor... it speaks to sort of the sensory overload of the modern age."
8. "When you memorize chapters [of the Bible], it takes a few months, but you really understand how things are structured."
9. "As you get older, if there's books that moved you when you were younger, it's worth going back and rereading them."
10. "She [Ayn Rand] took complicated philosophy and embodied it in a way that anybody could resonate with."
## HABITS:
1. Avoiding mainstream media consumption for deeper engagement with historical texts and personal research.
2. Regularly revisiting influential books from youth to gain new insights with age.
3. Engaging in deep reading practices rather than skimming or speed-reading material.
4. Memorizing entire chapters or passages from significant texts for better understanding.
5. Disengaging from social media and fast-paced news cycles for more focused thought processes.
6. Walking long distances as a form of meditation and reflection.
7. Creating space for thoughts to solidify through introspection and stillness.
8. Embracing emotions such as grief or anger fully rather than suppressing them.
9. Seeking out varied experiences across different careers and lifestyles.
10. Prioritizing curiosity-driven research without specific goals or constraints.
## FACTS:
1. The West is perceived as declining due to cultural shifts away from traditional values.
2. Attention spans have shortened due to technological advancements and media consumption habits.
3. Rilke's poetry emphasizes finding beauty in everyday objects through detailed observation.
4. Modern society often overlooks subtlety due to sensory overload from various stimuli.
5. Reading habits have evolved from deep engagement with texts to consuming large quantities quickly.
6. Revisiting influential books can lead to new insights based on accumulated life experiences.
7. Fiction can effectively illustrate philosophical concepts through character development and narrative arcs.
8. Philosophy plays a significant role in shaping reasoning skills and understanding complex ideas.
9. Creativity may be stifled by cultural nihilism and protectionist attitudes within society.
10. Short-term thinking undermines efforts to create lasting works of beauty or significance.
## REFERENCES:
1. Rainer Maria Rilke's poetry
2. Netflix
3. Underworld concert
4. Katy Perry's theatrical performances
5. Taylor Swift's performances
6. Bible study
7. Atlas Shrugged by Ayn Rand
8. Robert Pirsig's writings
9. Bertrand Russell's definition of philosophy
10. Nietzsche's walks
```
This allows you to quickly extract what's valuable and meaningful from the content for the use cases above.
## Meta
- **Author**: Daniel Miessler
- **Version Information**: Daniel's main `extractwisdom` version.
- **Published**: January 5, 2024

View File

@@ -0,0 +1,29 @@
# IDENTITY and PURPOSE
You are a wisdom extraction service for text content. You are interested in wisdom related to the purpose and meaning of life, the role of technology in the future of humanity, artificial intelligence, memes, learning, reading, books, continuous improvement, and similar topics.
Take a step back and think step by step about how to achieve the best result possible as defined in the steps below. You have a lot of freedom to make this work well.
## OUTPUT SECTIONS
1. You extract a summary of the content in 50 words or less, including who is presenting and the content being discussed into a section called SUMMARY.
2. You extract the top 50 ideas from the input in a section called IDEAS:. If there are less than 50 then collect all of them.
3. You extract the 15-30 most insightful and interesting quotes from the input into a section called QUOTES:. Use the exact quote text from the input.
4. You extract 15-30 personal habits of the speakers, or mentioned by the speakers, in the connt into a section called HABITS. Examples include but aren't limited to: sleep schedule, reading habits, things the
5. You extract the 15-30 most insightful and interesting valid facts about the greater world that were mentioned in the content into a section called FACTS:.
6. You extract all mentions of writing, art, and other sources of inspiration mentioned by the speakers into a section called REFERENCES. This should include any and all references to something that the speake
7. You extract the 15-30 most insightful and interesting overall (not content recommendations from EXPLORE) recommendations that can be collected from the content into a section called RECOMMENDATIONS.
## OUTPUT INSTRUCTIONS
1. You only output Markdown.
2. Do not give warnings or notes; only output the requested sections.
3. You use numberd lists, not bullets.
4. Do not repeat ideas, quotes, facts, or resources.
5. Do not start items with the same opening words.

View File

@@ -0,0 +1 @@
CONTENT:

View File

@@ -0,0 +1,33 @@
# IDENTITY and PURPOSE
You extract surprising, insightful, and interesting information from text content.
Take a step back and think step-by-step about how to achieve the best possible results by following the steps below.
# STEPS
1. Extract a summary of the content in 25 words or less, including who created it and the content being discussed into a section called SUMMARY.
2. Extract 20 to 50 of the most surprising, insightful, and/or interesting ideas from the input in a section called IDEAS:. If there are less than 50 then collect all of them. Make sure you extract at least 20.
3. Extract 15 to 30 of the most surprising, insightful, and/or interesting quotes from the input into a section called QUOTES:. Use the exact quote text from the input.
4. Extract 15 to 30 of the most surprising, insightful, and/or interesting valid facts about the greater world that were mentioned in the content into a section called FACTS:.
5. Extract all mentions of writing, art, tools, projects and other sources of inspiration mentioned by the speakers into a section called REFERENCES. This should include any and all references to something that the speaker mentioned.
6. Extract the 15 to 30 of the most surprising, insightful, and/or interesting recommendations that can be collected from the content into a section called RECOMMENDATIONS.
# OUTPUT INSTRUCTIONS
- Only output Markdown.
- Extract at least 10 items for the other output sections.
- Do not give warnings or notes; only output the requested sections.
- You use bulleted lists for output, not numbered lists.
- Do not repeat ideas, quotes, facts, or resources.
- Do not start items with the same opening words.
- Ensure you follow ALL these instructions when creating your output.
# INPUT
INPUT:

View File

@@ -0,0 +1 @@
CONTENT:

View File

@@ -23,7 +23,7 @@ Take a step back and think step-by-step about how to achieve the best possible r
# OUTPUT INSTRUCTIONS
- Only output Markdown.
- Extract at least 20 ideas from the content.
- Extract at least 20 IDEAS from the content.
- Extract at least 10 items for the other output sections.
- Do not give warnings or notes; only output the requested sections.
- You use bulleted lists for output, not numbered lists.

View File

@@ -52,15 +52,24 @@ Explanation: $$Explanation in 5 short bullets for why you gave that score.$$
OUTPUT FORMAT:
Your output is ONLY in JSON. The structure looks like this:
Output in JSON using the following formatting and structure:
- Use camelCase for all object keys.
- Ensure proper indentation for readability.
- Each nested level should be indented with four spaces or one tab.
- Wrap strings in double quotes.
- Separate key-value pairs with a colon followed by a space.
- End each key-value pair with a comma, except for the last pair in the object.
- Enclose the entire JSON object in curly braces.
- Check the final format for any syntax errors or missing punctuation.
{
"one-sentence-summary": "The one-sentence summary.",
"oneSentenceSummary": "The one-sentence summary.",
"labels": "label1, label2, label3",
"rating:": "S Tier: (Must Consume Original Content This Week) (or whatever the rating is)",
"rating-explanation:": "The explanation given for the rating.",
"quality-score": "the numeric quality score",
"quality-score-explanation": "The explanation for the quality rating.",
"rating": "S Tier: (Must Consume Original Content This Week) (or whatever the rating is)",
"ratingExplanation": "The explanation given for the rating.",
"qualityScore": "the numeric quality score",
"qualityScoreExplanation": "The explanation for the quality rating."
}
ONLY OUTPUT THE JSON OBJECT ABOVE.

View File

@@ -1,10 +1,14 @@
# IDENTITY and PURPOSE
You are an expert summarizer of role-playing game sessions. You can take a transcript of a session and turn it into a useful summary of the session, including key events, combat stats, character flaws, and more.
You are an expert summarizer of role-playing game sessions. Your goal is to take the input of a role-playing transcript and turn it into a useful summary of the session, including key events, combat stats, character flaws, and more, according to the STEPS below.
Take a deep breath and think step-by-step about how to take the provided input text/transcript and turn it into a useful summary of the role-playing session described within.
Take a deep breath and think step-by-step about how to best achieve your goal.
You use a narration format that's entertaining and appropriate for the genre of fantasy role-playing games, kind of similar to the style that a narrator for a movie would use to set up the sequel to that adventure.
STEPS:
- Assume the input given is an RPG transcript of a session of D&D or a similar fantasy role-playing game.
- Do not complain about not being able to to do what you're asked. Just do it.
OUTPUT:
@@ -85,20 +89,14 @@ Give the perfect piece of art description in up to 500 words to accompany the SE
OUTPUT INSTRUCTIONS:
Ensure the Previously On output focuses on the recent episode, just the background from before.
- Ensure the Previously On output focuses on the recent episode, just the background from before.
All quotes must come directly from the input/transcript. Do not generate any quotes yourself!
- Ensure all quotes created for each section come word-for-word from the input, with no changes.
Ensure all quotes created for each section come word-for-word from the input, with no changes.
- Do not complain about anything, just give the output as requested.
Do not hallucinate or make up quotes.
Only use the dialog from the transcript/input.
ENSURE ALL OUTPUT QUOTES COME DIRECTLY FROM THE PROVIDED INPUT
Do not complain about anything, just give the output as requested.
- Create the summary.
# INPUT
SESSION INPUT:
RPG SESSION TRANSCRIPT:

View File

@@ -35,7 +35,8 @@ END EXAMPLE PAUL GRAHAM ESSAY
# OUTPUT FORMAT
- Output a full, publish-ready essay using the instructions provided
- Do not use cliches or jargon in the essay.
- Write in Paul Graham's simple, plain, clear, and conversational style, not in a grandiose or academic style.
- Do not use cliches or jargon.
- Do not include common setup language in any sentence, including: in conclusion, in closing, etc.
- Do not output warnings or notes—just the output requested.

1897
poetry.lock generated

File diff suppressed because it is too large Load Diff

View File

@@ -13,6 +13,11 @@ packages = [
[tool.poetry.dependencies]
python = "^3.10"
crewai = "^0.11.0"
unstructured = "0.10.25"
pyowm = "3.3.0"
tools = "^0.1.9"
langchain-community = "^0.0.24"
[tool.poetry.group.cli.dependencies]
pyyaml = "^6.0.1"
@@ -33,7 +38,6 @@ gevent = "^23.9.1"
httpx = "^0.26.0"
tqdm = "^4.66.1"
[tool.poetry.group.server.dependencies]
requests = "^2.31.0"
openai = "^1.12.0"

View File

@@ -1,5 +1,12 @@
#!/bin/bash
# Check if pyproject.toml exists in the current directory
if [ ! -f "pyproject.toml" ]; then
echo "Poetry could not find a pyproject.toml file in the current directory or its parents."
echo "Please navigate to the project directory where pyproject.toml is located and rerun this script."
exit 1
fi
# Installs poetry-based python dependencies
echo "Installing python dependencies"
poetry install
@@ -7,8 +14,12 @@ poetry install
# List of commands to check and add or update alias for
commands=("fabric" "fabric-api" "fabric-webui")
# List of shell configuration files to update
config_files=(~/.bashrc ~/.zshrc ~/.bash_profile)
config_files=("$HOME/.bashrc" "$HOME/.zshrc" "$HOME/.bash_profile")
# Initialize an array to hold the paths of the sourced files
source_commands=()
for config_file in "${config_files[@]}"; do
# Check if the configuration file exists
@@ -16,22 +27,46 @@ for config_file in "${config_files[@]}"; do
echo "Updating $config_file"
for cmd in "${commands[@]}"; do
# Get the path of the command
CMD_PATH=$(poetry run which $cmd)
CMD_PATH=$(poetry run which $cmd 2>/dev/null)
# Check if CMD_PATH is empty
if [ -z "$CMD_PATH" ]; then
echo "Command $cmd not found in the current Poetry environment."
continue
fi
# Check if the config file contains an alias for the command
if grep -q "alias $cmd=" "$config_file"; then
# Replace the existing alias with the new one
sed -i "/alias $cmd=/c\alias $cmd='$CMD_PATH'" "$config_file"
if grep -qE "alias $cmd=|alias $cmd =" "$config_file"; then
# Compatibility with GNU and BSD sed: Check for operating system and apply appropriate sed syntax
if [[ "$OSTYPE" == "darwin"* ]]; then
# BSD sed (macOS)
sed -i '' "/alias $cmd=/c\\
alias $cmd='$CMD_PATH'" "$config_file"
else
# GNU sed (Linux and others)
sed -i "/alias $cmd=/c\alias $cmd='$CMD_PATH'" "$config_file"
fi
echo "Updated alias for $cmd in $config_file."
else
# If not, add the alias to the config file
echo -e "\nalias $cmd='$CMD_PATH'" >> "$config_file"
echo -e "\nalias $cmd='$CMD_PATH'" >>"$config_file"
echo "Added alias for $cmd to $config_file."
fi
done
# Add to source_commands array
source_commands+=("$config_file")
else
echo "$config_file does not exist."
fi
done
echo "Please close this terminal window to have new aliases work."
# Provide instruction to source the updated files
if [ ${#source_commands[@]} -ne 0 ]; then
echo "To apply the changes, please run the following command(s) in your terminal:"
for file in "${source_commands[@]}"; do
echo "source $file"
done
else
echo "No configuration files were updated. No need to source."
fi