Search This Blog

Sunday, November 26, 2017

Automated sound engineering software SFS WIN using python pywinauto

The program often bugs on function to save processed filed since this software has commercial version. There is one solution to try is to automate few steps with autohotkey and integrate with pyahk library, or place rande time.sleep method. The software may have internal time meter.

Monday, November 20, 2017

Bots for social media

In my pipeline many projects to facilitate tasks, to automate them to make business intelligence while scrapying data, process with pandas data frames. One high ROI project is completed - automation of computer tasks automation using Facebook suite. You can rent Facebook Suite, find customers and use industrial posting and other tasks platform to market. is python programmed bot, which post randomly in intervals safe for Facebook algorithms. Sounds perfect until you do it. Posting has to generate sales and good visibility to customers, and there should be many for low price to be worth doing it. With platform there is much manual job tasks which makes your life miserable.
Firstly, you have to collect much information of Facebook Groups - do not their purpose, members numbers, policies, if they are administered or not. To be profitable to you have to post to free groups, with high members numbers, to keep policies. To administer I programmed systematization tool and integrate data with pandas for sorting and update. Systematization program tool assigns members numbers to link and rules as well, therefore you can rearrange within one minute scraped from Facebook member profile Groups data to list - group name, members numbers and rules; and afterwards match  with links and names while merging files. The program is great process hundreds of targeted members data and get best plan for marketing with highest visibility at low costs.
Secondly, you have to post preferably up to 20 post and make pauses. So you have to delete first twenty groups. Manual job is required. There are tasks that can be done - collection and updating of moderated, approval, disabled for posting groups. It will help to save postings to improve visibility to potential customers and improve SEO and overall conversions of online business. These tasks have been automated with programs. Thirdly, you should often update your posting messages.
So I looked for  others of cheap traffic and audience source. Search "github python thumblr, twitter, pinterest, " shows many projects which are worth to test.  I will write about them later.

Sunday, November 19, 2017

Different text standards and encoding methods

While programming and making test you may have perfectly performing programs. But if you add data from various sources pro use libraries like pandas or other which encode or decode text, your programs using text file data may may not perform tasks as programmed. One problem I encountered that bot does not read see text data urls which passed through pandas encoding interpreter which in turn enodes everything to UTF-8, which I solved by downloading list using tolist bot. And another problem some url lines attaches each to other hindering programmed performance or making list of urls unreadable. Encoding and encoding methods detection is universal problem. It is useful to read for understanding What Every Programmer Absolutely, Positively Needs To Know About Encodings And Character Sets To Work With Text to understand how text works with encodings. To  detect  text encodings I use python chardect
Chardet: The Universal Character Encoding Detector which detects some encodings standards and some variants with detection confidenc. Today I tested I got:

C:\Users\ANTRAS>chardetect C:\Pohmeliy_FB\FB_post_to_groupsCR\lists\CRGroups.txt C:\Pohmeliy_FB\FB_post_to_groupsSILK\lists\SILKGroups.txt
C:\Pohmeliy_FB\FB_post_to_groupsCR\lists\CRGroups.txt: ascii with confidence 1.0
C:\Pohmeliy_FB\FB_post_to_groupsSILK\lists\SILKGroups.txt: ascii with confidence 1.0

So you you have to find one data source or decode and encode  to same byte order variant. About Byte order mark   you can read on .

Tuesday, November 14, 2017

PYAHK AutoHotKey via Python

AutoHotKey is a powerful task automator with a terrible scripting language built-in. Python is a powerful scripting language with old, unmaintained, and incomplete automation modules. Combining the two creates a powerful automation tool with a powerful scripting back-end with access to all the power of the Python standard library.
This is made possible by the amazing Python ctypes module and the AutoHotKey.dll project. Together they allow exchange of data between both scripting engines, and the execution of AHK commands from Python.
link PYAHK AutoHotKey via Python

Combining with pywinauto and SWAPY  automation tool and  PYAHK AutoHotKey via Python automated various software solutions can be programmed. Before testing I assume that cycle can be programmed using iterate range function, but for sound I have idea to analysis record speech and base iteration number to achieve necessary physics.

Sunday, November 12, 2017

Scrapy Crawling projects Search Github python scrapy

Useful projects for scrapy. I will test or use them for coding my tailor made. These crawlers do not include some pypi projects llike imagebot. 

scrapy-tdd 0.1.3 : Python Package Index - PyPI Helpers and examples to build Scrapy Crawlers in a test driven way. 
The project to be tested firstly.
For pypi scrapy crawlers I used search "pypi scrapy crawlers".

Today I code Pandas data import and updating and sorting by numbers as well strings and dropping repetitive lines. The day was successful. To be more productive I will program scrapy codes to collect information for freelancing.

Scrapy, a fast high-level web crawling & scraping framework for Python.
GitHub - geekan/scrapy-examples: Multifarious Scrapy examples .
scrapy-examples - Multifarious Scrapy examples. Spiders for alexa / amazon / douban / douyu / github / linkedin etc.

GitHub - scrapy/dirbot: Scrapy project to scrape public web directories.
dirbot - Scrapy project to scrape public web directories (educational) [DEPRECATED]

GitHub - scrapy/quotesbot: Quotebot for one bot.
This is a Scrapy project to scrape quotes from famous people from (github repo). This project is only meant for educational purposes.

GitHub - rmax/scrapy-redis: Redis-based components for Scrapy.
Redis-based components for Scrapy. Free software: MIT license; Documentation:

GitHub - scrapy/scrapely: A pure-python HTML screen-scraping library
A pure-python HTML screen-scraping library. Contribute to scrapely development by creating an account on GitHub.

GitHub - mjhea0/Scrapy-Samples: Scrapy examples crawling Craigslist
Scrapy examples crawling Craigslist. Contribute to Scrapy-Samples development by creating an account on GitHub.

GitHub - scrapinghub/portia: Visual scraping for Scrapy
Visual scraping for Scrapy. Contribute to portia development by creating an account on GitHub.

GitHub - edx/pa11ycrawler: Python crawler (using Scrapy) that uses.

pa11ycrawler - Python crawler (using Scrapy) that uses Pa11y to check accessibility of pages as it crawls.
GitHub - eloyz/reddit: .
2015-02-05 - Scrapy (Python Framework) Example using

GitHub - vinta/BlackWidow: Web crawler using Scrapy

Web crawler using Scrapy Install. $ sudo apt-get install python-dev libxml2-dev libxslt1-dev $ pip install -r requirements.txt.

GitHub - istresearch/scrapy-cluster: This Scrapy project uses Redis.
scrapy-cluster - This Scrapy project uses Redis and Kafka.

GitHub - scrapy/w3lib: Python library of web-related functions
Python library of web-related functions. Contribute to w3lib development by creating an account on GitHub.

GitHub - scrapy-plugins/scrapy-deltafetch: Scrapy spider middleware.

scrapy-deltafetch - Scrapy spider middleware to ignore requests to pages containing items ... DeltaFetch middleware depends on Python's bsddb3 package.

GitHub - scrapy/scrapyd: A service daemon to run Scrapy spiders
A service daemon to run Scrapy spiders. Scrapyd is a service for running Scrapy spiders.

Scrapy Plugins · GitHub
Scrapy spider middleware to ignore requests to pages containing items seen in previous crawls.

Web Scraping with Scrapy and MongoDB - Real Python

Deploy your Scrapy Spiders from GitHub – The Scrapinghub Blog
2017-04-19 - Up until now, your deployment process using Scrapy Cloud has probably ...
Scrapy Cloud's new GitHub integration will help you ensure that your.

python - Scrapy and github login - Stack Overflow
2016-11-26 - You shall try like this def parse(self, response): print "in parse function" yield FormRequest.from_response( response, ...

Running scrapy spider programmatically - Musings of a programmer
Please check the project on github. The Scrapy Spider : It is a python class in the scrapy framework that is responsible for fetching URLs and parsing the

scrapy-crawlera 1.2.4 : Python Package Index
Crawlera middleware for Scrapy. scrapy-crawlera 1.2.4 .Author: Raul Gallegos; Home Page:;

Webscraping Airbnb with scrapy – - Latest Posts
You can find the complete code here as github repo, feel free to fork, clone.

Scrapy Tutorial: Web Scraping Craigslist – Web Scraping with Python
Craigslist Scrapy Tutorial on GitHub - You can also find all the spiders we explained in this Python Scrapy tutorial on GitHub.

How do I change the data type of a pandas Series?