You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
209 lines
7.8 KiB
209 lines
7.8 KiB
from .utils import Standalone, Update, Setup, Alias, run_electron_app |
|
import argparse |
|
import sys |
|
import os |
|
|
|
|
|
script_directory = os.path.dirname(os.path.realpath(__file__)) |
|
|
|
|
|
def main(): |
|
parser = argparse.ArgumentParser( |
|
description="An open source framework for augmenting humans using AI." |
|
) |
|
parser.add_argument("--text", "-t", help="Text to extract summary from") |
|
parser.add_argument( |
|
"--copy", "-C", help="Copy the response to the clipboard", action="store_true" |
|
) |
|
parser.add_argument( |
|
'--agents', '-a', |
|
help="Use praisonAI to create an AI agent and then use it. ex: 'write me a movie script'", action="store_true" |
|
) |
|
|
|
parser.add_argument( |
|
"--output", |
|
"-o", |
|
help="Save the response to a file", |
|
nargs="?", |
|
const="analyzepaper.txt", |
|
default=None, |
|
) |
|
parser.add_argument('--session', '-S', |
|
help="Continue your previous conversation. Default is your previous conversation", nargs="?", const="default") |
|
parser.add_argument( |
|
'--clearsession', help="deletes indicated session. Use 'all' to delete all sessions") |
|
parser.add_argument('--sessionlog', help="View the log of a session") |
|
parser.add_argument( |
|
'--listsessions', help="List all sessions", action="store_true") |
|
parser.add_argument( |
|
"--gui", help="Use the GUI (Node and npm need to be installed)", action="store_true") |
|
parser.add_argument( |
|
"--stream", |
|
"-s", |
|
help="Use this option if you want to see the results in realtime. NOTE: You will not be able to pipe the output into another command.", |
|
action="store_true", |
|
) |
|
parser.add_argument( |
|
"--list", "-l", help="List available patterns", action="store_true" |
|
) |
|
parser.add_argument( |
|
'--temp', help="set the temperature for the model. Default is 0", default=0, type=float) |
|
parser.add_argument( |
|
'--top_p', help="set the top_p for the model. Default is 1", default=1, type=float) |
|
parser.add_argument( |
|
'--frequency_penalty', help="set the frequency penalty for the model. Default is 0.1", default=0.1, type=float) |
|
parser.add_argument( |
|
'--presence_penalty', help="set the presence penalty for the model. Default is 0.1", default=0.1, type=float) |
|
parser.add_argument( |
|
"--update", "-u", help="Update patterns. NOTE: This will revert the default model to gpt4-turbo. please run --changeDefaultModel to once again set default model", action="store_true") |
|
parser.add_argument("--pattern", "-p", help="The pattern (prompt) to use") |
|
parser.add_argument( |
|
"--setup", help="Set up your fabric instance", action="store_true" |
|
) |
|
parser.add_argument('--changeDefaultModel', |
|
help="Change the default model. For a list of available models, use the --listmodels flag.") |
|
|
|
parser.add_argument( |
|
"--model", "-m", help="Select the model to use" |
|
) |
|
parser.add_argument( |
|
"--listmodels", help="List all available models", action="store_true" |
|
) |
|
parser.add_argument('--remoteOllamaServer', |
|
help='The URL of the remote ollamaserver to use. ONLY USE THIS if you are using a local ollama server in an non-deault location or port') |
|
parser.add_argument('--context', '-c', |
|
help="Use Context file (context.md) to add context to your pattern", action="store_true") |
|
|
|
args = parser.parse_args() |
|
home_holder = os.path.expanduser("~") |
|
config = os.path.join(home_holder, ".config", "fabric") |
|
config_patterns_directory = os.path.join(config, "patterns") |
|
config_context = os.path.join(config, "context.md") |
|
env_file = os.path.join(config, ".env") |
|
if not os.path.exists(config): |
|
os.makedirs(config) |
|
if args.setup: |
|
Setup().run() |
|
Alias().execute() |
|
sys.exit() |
|
if not os.path.exists(env_file) or not os.path.exists(config_patterns_directory): |
|
print("Please run --setup to set up your API key and download patterns.") |
|
sys.exit() |
|
if not os.path.exists(config_patterns_directory): |
|
Update() |
|
Alias() |
|
sys.exit() |
|
if args.changeDefaultModel: |
|
Setup().default_model(args.changeDefaultModel) |
|
sys.exit() |
|
if args.gui: |
|
run_electron_app() |
|
sys.exit() |
|
if args.update: |
|
Update() |
|
Alias() |
|
sys.exit() |
|
if args.context: |
|
if not os.path.exists(os.path.join(config, "context.md")): |
|
print("Please create a context.md file in ~/.config/fabric") |
|
sys.exit() |
|
if args.agents: |
|
standalone = Standalone(args) |
|
text = "" # Initialize text variable |
|
# Check if an argument was provided to --agents |
|
if args.text: |
|
text = args.text |
|
else: |
|
text = standalone.get_cli_input() |
|
if text: |
|
standalone = Standalone(args) |
|
standalone.agents(text) |
|
sys.exit() |
|
if args.session: |
|
from .helper import Session |
|
session = Session() |
|
if args.session == "default": |
|
session_file = session.find_most_recent_file() |
|
if session_file is None: |
|
args.session = "default" |
|
else: |
|
args.session = session_file.split("/")[-1] |
|
if args.clearsession: |
|
from .helper import Session |
|
session = Session() |
|
session.clear_session(args.clearsession) |
|
if args.clearsession == "all": |
|
print(f"All sessions cleared") |
|
else: |
|
print(f"Session {args.clearsession} cleared") |
|
sys.exit() |
|
if args.sessionlog: |
|
from .helper import Session |
|
session = Session() |
|
print(session.session_log(args.sessionlog)) |
|
sys.exit() |
|
if args.listsessions: |
|
from .helper import Session |
|
session = Session() |
|
session.list_sessions() |
|
sys.exit() |
|
standalone = Standalone(args, args.pattern) |
|
if args.list: |
|
try: |
|
direct = sorted(os.listdir(config_patterns_directory)) |
|
for d in direct: |
|
print(d) |
|
sys.exit() |
|
except FileNotFoundError: |
|
print("No patterns found") |
|
sys.exit() |
|
if args.listmodels: |
|
gptmodels, localmodels, claudemodels = standalone.fetch_available_models() |
|
print("GPT Models:") |
|
for model in gptmodels: |
|
print(model) |
|
print("\nLocal Models:") |
|
for model in localmodels: |
|
print(model) |
|
print("\nClaude Models:") |
|
for model in claudemodels: |
|
print(model) |
|
sys.exit() |
|
if args.text is not None: |
|
text = args.text |
|
else: |
|
text = standalone.get_cli_input() |
|
if args.stream and not args.context: |
|
if args.remoteOllamaServer: |
|
standalone.streamMessage(text, host=args.remoteOllamaServer) |
|
else: |
|
standalone.streamMessage(text) |
|
sys.exit() |
|
if args.stream and args.context: |
|
with open(config_context, "r") as f: |
|
context = f.read() |
|
if args.remoteOllamaServer: |
|
standalone.streamMessage( |
|
text, context=context, host=args.remoteOllamaServer) |
|
else: |
|
standalone.streamMessage(text, context=context) |
|
sys.exit() |
|
elif args.context: |
|
with open(config_context, "r") as f: |
|
context = f.read() |
|
if args.remoteOllamaServer: |
|
standalone.sendMessage( |
|
text, context=context, host=args.remoteOllamaServer) |
|
else: |
|
standalone.sendMessage(text, context=context) |
|
sys.exit() |
|
else: |
|
if args.remoteOllamaServer: |
|
standalone.sendMessage(text, host=args.remoteOllamaServer) |
|
else: |
|
standalone.sendMessage(text) |
|
sys.exit() |
|
|
|
|
|
if __name__ == "__main__": |
|
main()
|
|
|