main
Kye 2 years ago
parent d07cbbca3e
commit 8b3dc2b9a4

@ -480,16 +480,16 @@ class Terminal(BaseToolSet):
return output return output
if __name__ == "__main__": # if __name__ == "__main__":
import time # import time
o = Terminal().execute( # o = Terminal().execute(
"sleep 1; echo 1; sleep 2; echo 2; sleep 3; echo 3; sleep 10;", # "sleep 1; echo 1; sleep 2; echo 2; sleep 3; echo 3; sleep 10;",
lambda: ("", None), # lambda: ("", None),
) # )
print(o) # print(o)
time.sleep(10) # see if timer has reset # time.sleep(10) # see if timer has reset
###################=> EDITOR/VERIFY ###################=> EDITOR/VERIFY
@ -734,9 +734,9 @@ class CodeReader:
return SummaryCommand.from_str(command).execute() return SummaryCommand.from_str(command).execute()
if __name__ == "__main__": # if __name__ == "__main__":
summary = CodeReader.summary("read.py|1|class ReadCommand:") # summary = CodeReader.summary("read.py|1|class ReadCommand:")
print(summary) # print(summary)
#============================> EDITOR/READ.PY END #============================> EDITOR/READ.PY END
@ -908,40 +908,40 @@ class CodePatcher:
return written, deleted return written, deleted
if __name__ == "__main__": # if __name__ == "__main__":
commands = """test.py|2,1|2,1|from bs4 import BeautifulSoup # commands = """test.py|2,1|2,1|from bs4 import BeautifulSoup
---~~~+++===+++~~~--- # ---~~~+++===+++~~~---
test.py|5,5|5,33|html = requests.get(url).text # test.py|5,5|5,33|html = requests.get(url).text
soup = BeautifulSoup(html, "html.parser") # soup = BeautifulSoup(html, "html.parser")
news_results = soup.find_all("div", class_="BNeawe vvjwJb AP7Wnd") # news_results = soup.find_all("div", class_="BNeawe vvjwJb AP7Wnd")
---~~~+++===+++~~~--- # ---~~~+++===+++~~~---
test.py|7,5|9,13|news_titles = [] # test.py|7,5|9,13|news_titles = []
for result in news_results: # for result in news_results:
news_titles # news_titles
---~~~+++===+++~~~--- # ---~~~+++===+++~~~---
test.py|11,16|11,16|_titles # test.py|11,16|11,16|_titles
""" # """
example = """import requests # example = """import requests
def crawl_news(keyword): # def crawl_news(keyword):
url = f"https://www.google.com/search?q={keyword}+news" # url = f"https://www.google.com/search?q={keyword}+news"
response = requests.get(url) # response = requests.get(url)
news = [] # news = []
for result in response: # for result in response:
news.append(result.text) # news.append(result.text)
return news # return news
""" # """
testfile = "test.py" # testfile = "test.py"
with open(testfile, "w") as f: # with open(testfile, "w") as f:
f.write(example) # f.write(example)
patcher = CodePatcher() # patcher = CodePatcher()
written, deleted = patcher.patch(commands) # written, deleted = patcher.patch(commands)
print(f"written: {written}, deleted: {deleted}") # print(f"written: {written}, deleted: {deleted}")
####################### => EDITOR/PATCH.PY ####################### => EDITOR/PATCH.PY

Loading…
Cancel
Save