%PDF- %PDF-
Mini Shell

Mini Shell

Direktori : /usr/share/doc/python3-mechanize/html/_sources/
Upload File :
Create Path :
Current File : //usr/share/doc/python3-mechanize/html/_sources/advanced.rst.txt

Advanced topics
==================

.. _threading:

Thread safety
---------------

The global :func:`mechanize.urlopen()` and :func:`mechanize.urlretrieve()` functions are
thread safe. However, mechanize browser instances **are not** thread safe. If
you want to use a mechanize Browser instance in multiple threads, clone it,
using `copy.copy(browser_object)` method. The clone will share the same,
thread safe cookie jar, and have the same settings/handlers as the original,
but all other state is not shared, making the clone safe to use in a different
thread.

Using custom CA certificates
-------------------------------

mechanize supports the same mechanism for using custom CA certificates as
python >= 2.7.9. To change the certificates a mechanize browser instance uses,
call the :meth:`mechanize.Browser.set_ca_data()` method on it. 

.. _debugging:

Debugging
--------------

Hints for debugging programs that use mechanize.

.. _cookies:

Cookies
^^^^^^^^^^

A common mistake is to use :func:`mechanize.urlopen()`, *and* the
`.extract_cookies()` and `.add_cookie_header()` methods on a cookie object
themselves.  If you use `mechanize.urlopen()` (or `OpenerDirector.open()`), the
module handles extraction and adding of cookies by itself, so you should not
call `.extract_cookies()` or `.add_cookie_header()`.

Are you sure the server is sending you any cookies in the first place?  Maybe
the server is keeping track of state in some other way (`HIDDEN` HTML form
entries (possibly in a separate page referenced by a frame), URL-encoded
session keys, IP address, HTTP `Referer` headers)?  Perhaps some embedded
script in the HTML is setting cookies (see below)?  Turn on :ref:`logging`.

When you `.save()` to or `.load()`/`.revert()` from a file, single-session
cookies will expire unless you explicitly request otherwise with the
`ignore_discard` argument.  This may be your problem if you find cookies are
going away after saving and loading.

.. code-block:: python

    import mechanize
    cj = mechanize.LWPCookieJar()
    opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cj))
    mechanize.install_opener(opener)
    r = mechanize.urlopen("http://foobar.com/")
    cj.save("/some/file", ignore_discard=True, ignore_expires=True)

JavaScript code can set cookies; mechanize does not support this.  See
:ref:`jsfaq`.


General
^^^^^^^^^

Enable :ref:`logging`.

Sometimes, a server wants particular HTTP headers set to the values it expects.
For example, the `User-Agent` header may need to be set
(:meth:`mechanize.Browser.set_header()`) to a value like that of a popular
browser.

Check that the browser is able to do manually what you're trying to achieve
programmatically.  Make sure that what you do manually is *exactly* the same as
what you're trying to do from Python -- you may simply be hitting a server bug
that only gets revealed if you view pages in a particular order, for example.

Try comparing the headers and data that your program sends with those that a
browser sends.  Often this will give you the clue you need.  You can use
the developer tools in any browser to see exactly what the browser sends and
receives.

If nothing is obviously wrong with the requests your program is sending and
you're out of ideas, you can reliably locate the problem by copying the headers
that a browser sends, and then changing headers until your program stops
working again.  Temporarily switch to explicitly sending individual HTTP
headers (by calling `.add_header()`, or by using `httplib` directly).  Start by
sending exactly the headers that Firefox or Chrome send.  You may need to make sure
that a valid session ID is sent -- the one you got from your browser may no
longer be valid.  If that works, you can begin the tedious process of changing
your headers and data until they match what your original code was sending.
You should end up with a minimal set of changes.  If you think that reveals a
bug in mechanize, please report it.


.. _logging:

Logging
^^^^^^^^^

To enable logging to stdout:

.. code-block:: python
    
    import sys, logging
    logger = logging.getLogger("mechanize")
    logger.addHandler(logging.StreamHandler(sys.stdout))
    logger.setLevel(logging.DEBUG)

You can reduce the amount of information shown by setting the level to
`logging.INFO` instead of `logging.DEBUG`, or by only enabling logging for one
of the following logger names instead of `"mechanize"`:

  * `"mechanize"`: Everything.

  * `"mechanize.cookies"`: Why particular cookies are accepted or rejected and why
    they are or are not returned.  Requires logging enabled at the `DEBUG` level.

  * `"mechanize.http_responses"`: HTTP response body data.

  * `"mechanize.http_redirects"`: HTTP redirect information.


HTTP headers
^^^^^^^^^^^^^

An example showing how to enable printing of HTTP headers to stdout, logging of
HTTP response bodies, and logging of information about redirections:

.. code-block:: python

    import sys, logging
    import mechanize

    logger = logging.getLogger("mechanize")
    logger.addHandler(logging.StreamHandler(sys.stdout))
    logger.setLevel(logging.DEBUG)

    browser = mechanize.Browser()
    browser.set_debug_http(True)
    browser.set_debug_responses(True)
    browser.set_debug_redirects(True)
    response = browser.open("http://python.org/")

Alternatively, you can examine request and response objects to see what's going
on.  Note that requests may involve "sub-requests" in cases such as
redirection, in which case you will not see everything that's going on just by
examining the original request and final response.  

Zerion Mini Shell 1.0