Commit Graph

6 Commits

Author SHA1 Message Date
Christian Clauss
5df8aec66c
GitHub Action formats our code with psf/black (#1569)
* GitHub Action formats our code with psf/black

@poyea Your review please.

* fixup! Format Python code with psf/black push
2019-11-14 19:59:43 +01:00
Sarath Kaul
dd7d2fa270 Panagram Script Added (#1564)
* Python Program that fetches top trending news

* Python Program that fetches top trending news

* Revisions in Fetch BBC News

* psf/black Changes

* Python Program to send slack message to a channel

* Slack Message Revision Changes

* Python Program to check Palindrome String

* Doctest Added

* Python Program to check whether a String is Panagram or not

* Python Program to check whether a String is Panagram or not

* Check Panagram Script Added

* Panagram Script Added

* Anagram Script Changes

* Anagram Alphabet Check Added

* Python Program to fetch github info
2019-11-14 11:22:07 +01:00
Sarath Kaul
b7e37a856f Added a new Python script and some changes in existing one (#1560)
* Python Program that fetches top trending news

* Python Program that fetches top trending news

* Revisions in Fetch BBC News

* psf/black Changes

* Python Program to send slack message to a channel

* Slack Message Revision Changes
2019-11-12 12:11:54 +01:00
Sarath Kaul
5ac4391420 Python Program that fetches top trending news (#1559)
* Python Program that fetches top trending news

* Python Program that fetches top trending news

* Revisions in Fetch BBC News
2019-11-12 09:27:38 +01:00
Orkun İncili
8b52e44230 Python program that listing top 'n' movie in imdb (#1477)
* Python program that listing top 'n' movie in imdb

* Update get_imdbtop.py
2019-10-27 21:48:38 +01:00
kunal kumar barman
43f99e56c9 Python program that surfs 3 site at a time (#1389)
* Python program that scrufs 3 site at a time

add input in the compiling time  like --  python3 project1.py (man)

* Update project1.py

* noqa: F401 and reformat with black

* Rename project1.py to web_programming/crawl_google_results.py

* Add beautifulsoup4 to requirements.txt

* Add fake_useragent to requirements.txt

* Update crawl_google_results.py

* headers={"UserAgent": UserAgent().random}

* html.parser, not lxml

* link, not links
2019-10-18 23:30:52 +02:00