User:Kaori Emora/user-config.py

This file is found in the  directory. See also: http://www.mediawiki.org/wiki/Manual:Pywikipediabot/Use_on_non-WMF_wikis
 * 1) -*- coding: utf-8  -*-


 * 1) This is an automatically generated file. You can find more configuration
 * 2) parameters in 'config.py' file.

family = 'lolwiki'
 * 1) The family of sites we are working on. wikipedia.py will import
 * 2) families/xxx_family.py so if you want to change this variable,
 * 3) you need to write such a file.

mylang = 'en'
 * 1) The language code of the site we're working on.

usernames['lolwiki']['en'] = u'Kaori Emora'
 * 1) The dictionary usernames should contain a username for each site where you
 * 2) have a bot account.


 * 1) LOGFILE SETTINGS ##############

log = ['interwiki']
 * 1) Defines for which scripts a logfile should be enabled. Logfiles will be
 * 2) saved in the 'logs' subdirectory.
 * 3) Example:
 * 4)     log = ['interwiki', 'weblinkchecker', 'table2wiki']
 * 5) It is also possible to enable logging for all scripts, using this line:
 * 6)     log = ['*']
 * 7) To disable all logging, use this:
 * 8)     log = []
 * 9) Per default, logging of interwiki.py is enabled because its logfiles can
 * 10) be used to generate so-called warnfiles.
 * 11) This setting can be overridden by the -log or -nolog command-line arguments.


 * 1) INTERWIKI SETTINGS ##############

interwiki_backlink = True
 * 1) Should interwiki.py report warnings for missing links between foreign
 * 2) languages?

interwiki_shownew = True
 * 1) Should interwiki.py display every new link it discovers?

interwiki_graph = False
 * 1) Should interwiki.py output a graph PNG file on conflicts?
 * 2) You need pydot for this: http://dkbza.org/pydot.html

interwiki_min_subjects = 100
 * 1) Specifies that the robot should process that amount of subjects at a time,
 * 2) only starting to load new pages in the original language when the total
 * 3) falls below that number. Default is to process (at least) 100 subjects at
 * 4) once.

interwiki_graph_formats = ['png']
 * 1) If interwiki graphs are enabled, which format(s) should be used?
 * 2) Supported formats include png, jpg, ps, and svg. See:
 * 3) http://www.graphviz.org/doc/info/output.html
 * 4) If you want to also dump the dot files, you can use this in your
 * 5) user-config.py:
 * 6) interwiki_graph_formats = ['dot', 'png']
 * 7) If you need a PNG image with an HTML image map, use this:
 * 8) interwiki_graph_formats = ['png', 'cmap']
 * 9) If you only need SVG images, use:
 * 10) interwiki_graph_formats = ['svg']

interwiki_graph_url = None
 * 1) You can post the contents of your autonomous_problems.dat to the wiki,
 * 2) e.g. to http://de.wikipedia.org/wiki/Wikipedia:Interwiki-Konflikte.
 * 3) This allows others to assist you in resolving interwiki problems.
 * 4) To help these people, you can upload the interwiki graphs to your
 * 5) webspace somewhere. Set the base URL here, e.g.:
 * 6) 'http://www.example.org/~yourname/interwiki-graphs/'

without_interwiki = False
 * 1) Save file with local articles without interwikis.

interwiki_contents_on_disk = False
 * 1) Experimental feature:
 * 2) Store the page contents on disk (/cache/ directory) instead of loading
 * 3) them in RAM.


 * 1) SOLVE_DISAMBIGUATION SETTINGS ############
 * 2) Set disambiguation_comment[FAMILY][LANG] to a non-empty string to override
 * 3) the default edit comment for the solve_disambiguation bot.
 * 4) Use %s to represent the name of the disambiguation page being treated.
 * 5) Example:
 * 6) disambiguation_comment['wikipedia']['en'] = \
 * 7)    "Robot-assisted disambiguation (you can help!): %s"
 * 1) disambiguation_comment['wikipedia']['en'] = \
 * 2)    "Robot-assisted disambiguation (you can help!): %s"

disambiguation_comment = {} sort_ignore_case = False

upload_to_commons = False
 * 1) IMAGE RELATED SETTINGS ##############
 * 2) If you set this to True, images will be uploaded to Wikimedia
 * 3) Commons by default.


 * 1) TABLE CONVERSION BOT SETTINGS ##############

splitLongParagraphs = False deIndentTables = True table2wikiAskOnlyWarnings = True table2wikiSkipWarnings = False
 * 1) will split long paragraphs for better reading the source.
 * 2) only table2wiki.py use it by now
 * 1) sometimes HTML-tables are indented for better reading.
 * 2) That can do very ugly results.
 * 1) table2wiki.py works quite stable, so you might switch to True


 * 1) WEBLINK CHECKER SETTINGS ##############

max_external_links = 50
 * 1) How many external links should weblinkchecker.py check at the same time?
 * 2) If you have a fast connection, you might want to increase this number so
 * 3) that slow servers won't slow you down.

report_dead_links_on_talk = False

db_hostname = 'localhost' db_username = 'wikiuser' db_password = ''
 * 1) DATABASE SETTINGS ##############


 * 1) SEARCH ENGINE SETTINGS ##############

google_key = ''
 * 1) Some scripts allow querying Google via the Google Web API. To use this feature,
 * 2) you must install the pyGoogle module from http://pygoogle.sf.net/ and have a
 * 3) Google Web API license key. Note that Google doesn't give out license keys
 * 4) anymore.
 * 5) Google web API is obsoleted for long time, now we can use Google AJAX Search API,
 * 6) You can signup an API key from http://code.google.com/apis/ajaxsearch/signup.html.
 * 1) You can signup an API key from http://code.google.com/apis/ajaxsearch/signup.html.

google_api_refer = ''
 * 1) using Google AJAX Search API, it requires the referer website, this variable saves the referer web address
 * 2) when you sign up with the key.

yahoo_appid = ''
 * 1) Some scripts allow using the Yahoo! Search Web Services. To use this feature,
 * 2) you must install the pYsearch module from http://pysearch.sourceforge.net/
 * 3) and get a Yahoo AppID from http://developer.yahoo.com

msn_appid = ''
 * 1) To use Windows Live Search web service you must get an AppID from
 * 2) http://search.msn.com/developer

flickr = { 'api_key': u'', # Provide your key! 'review': False, # Do we use automatically make our uploads reviewed? 'reviewer': u'', # If so, under what reviewer name? }
 * 1) Using the Flickr api

panoramio = { 'review': False, # Do we use automatically make our uploads reviewed? 'reviewer': u'', # If so, under what reviewer name? }
 * 1) Using the Panoramio api

proxy = { 'host': None, 'auth': None, }
 * 1) for all connections: proxy handle
 * 2) to use it, proxy['host'] have to support HTTP and include port number (e.g. localhost:8080)
 * 3) if the proxy server needs authentication, set ('ID', 'PASSWORD') to proxy['auth'].


 * 1) COPYRIGHT SETTINGS ##############

copyright_google = True copyright_yahoo = True copyright_msn = False
 * 1) Enable/disable search engine in copyright.py script

copyright_check_in_source_google = False copyright_check_in_source_yahoo = False copyright_check_in_source_msn = False
 * 1) Perform a deep check, loading URLs to search if 'Wikipedia' is present.
 * 2) This may be useful to increase the number of correct results. If you haven't
 * 3) a fast connection, you might want to keep them disabled.

copyright_check_in_source_section_names = False
 * 1) Web pages may contain a Wikipedia text without the word 'Wikipedia' but with
 * 2) the typical '[edit]' tag as a result of a copy & paste procedure. You want
 * 3) no report for this kind of URLs, even if they are copyright violations.
 * 4) However, when enabled, these URLs are logged in a file.

copyright_max_query_for_page = 25
 * 1) Limit number of queries for page.

copyright_skip_query = 0
 * 1) Skip a specified number of queries

copyright_connection_tries = 10
 * 1) Number of attempts on connection error.

copyright_exceeded_in_queries = 2 copyright_exceeded_in_queries_sleep_hours = 6
 * 1) Behavior if an exceeded error occur.
 * 2) Possibilities:
 * 3)    0 = None
 * 4)    1 = Disable search engine
 * 5)    2 = Sleep (default)
 * 6)    3 = Stop
 * 1)    2 = Sleep (default)
 * 2)    3 = Stop

copyright_show_date = True
 * 1) Append last modified date of URL to script result

copyright_show_length = True
 * 1) Append length of URL to script result

copyright_economize_query = True
 * 1) By default the script tries to identify and skip text that contains a large
 * 2) comma separated list or only numbers. But sometimes that might be the
 * 3) only part unmodified of a slightly edited and not otherwise reported
 * 4) copyright violation. You can disable this feature to try to increase the
 * 5) number of results.


 * 1) FURTHER SETTINGS ##############

use_diskcache = False
 * 1) Use the experimental disk cache to prevent huge memory usage

retry_on_fail = True
 * 1) Retry loading a page on failure (back off 1 minute, 2 minutes, 4 minutes
 * 2) up to 30 minutes)

actions_to_block = ['edit', 'watch', 'move', 'delete', 'undelete', 'protect', 'emailuser', 'submit']   # 'submit' is for non-API
 * 1) Defines what actions the bots are NOT allowed to do (e.g. 'edit') on wikipedia
 * 2) servers. Allows simulation runs of bots to be carried out without changing any
 * 3) page on the server side. This setting may be overridden in user_config.py.

max_queue_size = 64
 * 1) How many pages should be put to a queue in asynchroneous mode.
 * 2) If maxsize is <= 0, the queue size is infinite.
 * 3) Increasing this value will increase memory space but could speed up
 * 4) processing. As higher this value this effect will decrease.

line_separator = LS = u'\r\n'
 * 1) Define the line separator. Pages retrieved via API have "\n" whereas
 * 2) pages fetched from screen (mostly) have "\r\n". Interwiki and category
 * 3) separator settings in family files should use multiplied of this.
 * 4) LS is a shortcut alias.