libcurl-tutorial(3) - phpMan

Command: man perldoc info search(apropos)  


libcurl-tutorial(3)                    libcurl programming                    libcurl-tutorial(3)



NAME
       libcurl-tutorial - libcurl programming tutorial

Objective
       This  document  attempts  to  describe the general principles and some basic approaches to
       consider when programming with libcurl. The text will focus mainly on the C interface  but
       might  apply  fairly  well  on  other  interfaces as well as they usually follow the C one
       pretty closely.

       This document will refer to 'the user' as the person writing the  source  code  that  uses
       libcurl.  That  would probably be you or someone in your position.  What will be generally
       referred to as 'the program' will be the collected source code  that  you  write  that  is
       using  libcurl for transfers. The program is outside libcurl and libcurl is outside of the
       program.

       To get more details on all options and functions described herein, please refer  to  their
       respective man pages.


Building
       There  are  many different ways to build C programs. This chapter will assume a UNIX-style
       build process. If you use a different build system, you can still read this to get general
       information that may apply to your environment as well.

       Compiling the Program
              Your  compiler  needs  to know where the libcurl headers are located. Therefore you
              must set your compiler's include path to point to the directory where you installed
              them. The 'curl-config'[3] tool can be used to get this information:

              $ curl-config --cflags


       Linking the Program with libcurl
              When  having  compiled  the program, you need to link your object files to create a
              single executable. For that to succeed, you need to link with libcurl and  possibly
              also  with  other  libraries  that  libcurl  itself  depends  on.  Like the OpenSSL
              libraries, but even some standard OS libraries may be needed on the  command  line.
              To  figure  out  which flags to use, once again the 'curl-config' tool comes to the
              rescue:

              $ curl-config --libs


       SSL or Not
              libcurl can be built and customized in many ways. One of  the  things  that  varies
              from  different  libraries  and builds is the support for SSL-based transfers, like
              HTTPS and FTPS. If a supported SSL library was  detected  properly  at  build-time,
              libcurl  will  be built with SSL support. To figure out if an installed libcurl has
              been built with SSL support enabled, use 'curl-config' like this:

              $ curl-config --feature

              And if SSL is supported, the keyword 'SSL' will  be  written  to  stdout,  possibly
              together  with a few other features that could be either on or off on for different
              libcurls.

              See also the "Features libcurl Provides" further down.

       autoconf macro
              When you write your configure script to detect libcurl and setup variables  accord-
              ingly,  we  offer a prewritten macro that probably does everything you need in this
              area. See docs/libcurl/libcurl.m4 file - it includes docs on how to use it.


Portable Code in a Portable World
       The people behind libcurl have put a considerable effort to make libcurl work on  a  large
       amount of different operating systems and environments.

       You  program  libcurl  the  same way on all platforms that libcurl runs on. There are only
       very few minor considerations that differ. If you just make sure to write your code porta-
       ble  enough,  you may very well create yourself a very portable program. libcurl shouldn't
       stop you from that.


Global Preparation
       The program must initialize some of the libcurl  functionality  globally.  That  means  it
       should  be done exactly once, no matter how many times you intend to use the library. Once
       for your program's entire life time. This is done using

        curl_global_init()

       and it takes one parameter which is a bit pattern that tells libcurl what  to  initialize.
       Using CURL_GLOBAL_ALL will make it initialize all known internal sub modules, and might be
       a good default option. The current two bits that are specified are:

              CURL_GLOBAL_WIN32
                     which only does anything  on  Windows  machines.  When  used  on  a  Windows
                     machine,  it'll make libcurl initialize the win32 socket stuff. Without hav-
                     ing that initialized properly, your program cannot use sockets properly. You
                     should  only  do  this once for each application, so if your program already
                     does this or of another library in use does it, you should not tell  libcurl
                     to do this as well.

              CURL_GLOBAL_SSL
                     which  only  does  anything  on  libcurls compiled and built SSL-enabled. On
                     these systems, this will make libcurl initialize the  SSL  library  properly
                     for  this  application. This only needs to be done once for each application
                     so if your program or another library already does this, this bit should not
                     be needed.

       libcurl has a default protection mechanism that detects if curl_global_init(3) hasn't been
       called by the time curl_easy_perform(3) is called and if that is the  case,  libcurl  runs
       the  function itself with a guessed bit pattern. Please note that depending solely on this
       is not considered nice nor very good.

       When the program no longer uses libcurl, it should call curl_global_cleanup(3),  which  is
       the  opposite  of  the  init  call. It will then do the reversed operations to cleanup the
       resources the curl_global_init(3) call initialized.

       Repeated calls to curl_global_init(3) and curl_global_cleanup(3) should be  avoided.  They
       should only be called once each.


Features libcurl Provides
       It  is  considered  best-practice to determine libcurl features at run-time rather than at
       build-time (if possible of course). By calling curl_version_info(3) and checking  out  the
       details  of  the  returned  struct, your program can figure out exactly what the currently
       running libcurl supports.


Handle the Easy libcurl
       libcurl first introduced the so called easy interface. All operations in the  easy  inter-
       face are prefixed with 'curl_easy'.

       Recent libcurl versions also offer the multi interface. More about that interface, what it
       is targeted for and how to use it is detailed in a  separate  chapter  further  down.  You
       still  need  to understand the easy interface first, so please continue reading for better
       understanding.

       To use the easy interface, you must first create yourself an easy  handle.  You  need  one
       handle for each easy session you want to perform. Basically, you should use one handle for
       every thread you plan to use for transferring. You must never share  the  same  handle  in
       multiple threads.

       Get an easy handle with

        easyhandle = curl_easy_init();

       It  returns  an easy handle. Using that you proceed to the next step: setting up your pre-
       ferred actions. A handle is just a logic entity for the upcoming  transfer  or  series  of
       transfers.

       You set properties and options for this handle using curl_easy_setopt(3). They control how
       the subsequent transfer or transfers will be made. Options remain set in the handle  until
       set  again  to something different. Alas, multiple requests using the same handle will use
       the same options.

       Many of the options you set in libcurl are "strings", pointers to data terminated  with  a
       zero  byte.  When  you set strings with curl_easy_setopt(3), libcurl makes its own copy so
       that they don't need to be kept around in your application after being set[4].

       One of the most basic properties to set in the handle is the URL. You set  your  preferred
       URL to transfer with CURLOPT_URL in a manner similar to:

        curl_easy_setopt(handle, CURLOPT_URL, "http://domain.com/");

       Let's  assume  for  a  while  that you want to receive data as the URL identifies a remote
       resource you want to get here. Since you write a  sort  of  application  that  needs  this
       transfer,  I  assume that you would like to get the data passed to you directly instead of
       simply getting it passed to stdout. So, you write your own function that matches this pro-
       totype:

        size_t write_data(void *buffer, size_t size, size_t nmemb, void *userp);

       You tell libcurl to pass all data to this function by issuing a function similar to this:

        curl_easy_setopt(easyhandle, CURLOPT_WRITEFUNCTION, write_data);

       You  can  control  what data your callback function gets in the fourth argument by setting
       another property:

        curl_easy_setopt(easyhandle, CURLOPT_WRITEDATA, &internal_struct);

       Using that property, you can easily pass local data between your application and the func-
       tion  that gets invoked by libcurl. libcurl itself won't touch the data you pass with CUR-
       LOPT_WRITEDATA.

       libcurl offers its own default internal callback that will take care of the  data  if  you
       don't set the callback with CURLOPT_WRITEFUNCTION. It will then simply output the received
       data to stdout. You can have the default callback write the data to a different file  han-
       dle by passing a 'FILE *' to a file opened for writing with the CURLOPT_WRITEDATA option.

       Now,  we  need  to take a step back and have a deep breath. Here's one of those rare plat-
       form-dependent nitpicks. Did you spot it? On some platforms[2], libcurl won't be  able  to
       operate  on files opened by the program. Thus, if you use the default callback and pass in
       an open file with CURLOPT_WRITEDATA, it will crash. You should  therefore  avoid  this  to
       make your program run fine virtually everywhere.

       (CURLOPT_WRITEDATA  was  formerly  known as CURLOPT_FILE. Both names still work and do the
       same thing).

       If you're using libcurl as a win32 DLL, you MUST use the CURLOPT_WRITEFUNCTION if you  set
       CURLOPT_WRITEDATA - or you will experience crashes.

       There  are  of  course  many more options you can set, and we'll get back to a few of them
       later. Let's instead continue to the actual transfer:

        success = curl_easy_perform(easyhandle);

       curl_easy_perform(3) will connect to the  remote  site,  do  the  necessary  commands  and
       receive  the transfer. Whenever it receives data, it calls the callback function we previ-
       ously set. The function may get one byte at a time, or it may get many kilobytes at  once.
       libcurl  delivers  as much as possible as often as possible. Your callback function should
       return the number of bytes it "took care of". If that is not  the  exact  same  amount  of
       bytes  that  was  passed  to it, libcurl will abort the operation and return with an error
       code.

       When the transfer is complete, the function returns a return code that informs you  if  it
       succeeded  in  its  mission or not. If a return code isn't enough for you, you can use the
       CURLOPT_ERRORBUFFER to point libcurl to a buffer of yours where it'll store a human  read-
       able error message as well.

       If you then want to transfer another file, the handle is ready to be used again. Mind you,
       it is even preferred that you re-use an existing handle if  you  intend  to  make  another
       transfer. libcurl will then attempt to re-use the previous connection.

       For  some  protocols,  downloading a file can involve a complicated process of logging in,
       setting the transfer mode, changing the current directory  and  finally  transferring  the
       file  data. libcurl takes care of all that complication for you. Given simply the URL to a
       file, libcurl will take care of all the details needed to get  the  file  moved  from  one
       machine to another.


Multi-threading Issues
       The  first  basic rule is that you must never simultaneously share a libcurl handle (be it
       easy or multi or whatever) between multiple threads. Only use one handle in one thread  at
       any  time.  You can pass the handles around among threads, but you must never use a single
       handle from more than one thread at any given time.

       libcurl is completely thread safe, except for two issues: signals  and  SSL/TLS  handlers.
       Signals  are used for timing out name resolves (during DNS lookup) - when built without c-
       ares support and not on Windows.

       If you are accessing HTTPS or FTPS URLs in a multi-threaded manner, you are then of course
       using  the  underlying  SSL  library  multi-threaded  and  those libs might have their own
       requirements on this issue. Basically, you need to provide one or two functions  to  allow
       it to function properly. For all details, see this:

       OpenSSL

        http://www.openssl.org/docs/crypto/threads.html#DESCRIPTION

       GnuTLS

        http://www.gnu.org/software/gnutls/manual/html_node/Multi_002dthreaded-applications.html

       NSS

        is claimed to be thread-safe already without anything required.

       PolarSSL

        Required actions unknown.

       yassl

        Required actions unknown.

       axTLS

        Required actions unknown.

       When  using  multiple threads you should set the CURLOPT_NOSIGNAL option to 1 for all han-
       dles. Everything will or might work fine except that timeouts are not honored  during  the
       DNS  lookup - which you can work around by building libcurl with c-ares support. c-ares is
       a library that provides asynchronous name resolves. On some platforms, libcurl simply will
       not function properly multi-threaded unless this option is set.

       Also, note that CURLOPT_DNS_USE_GLOBAL_CACHE is not thread-safe.


When It Doesn't Work
       There will always be times when the transfer fails for some reason. You might have set the
       wrong libcurl option or misunderstood what the libcurl option actually does, or the remote
       server might return non-standard replies that confuse the library which then confuses your
       program.

       There's one golden rule when these things occur: set  the  CURLOPT_VERBOSE  option  to  1.
       It'll  cause  the  library to spew out the entire protocol details it sends, some internal
       info and some received protocol data as well (especially when using FTP). If you're  using
       HTTP,  adding  the  headers  in the received output to study is also a clever way to get a
       better understanding why the server behaves the way it does. Include headers in the normal
       body output with CURLOPT_HEADER set 1.

       Of  course,  there  are  bugs  left. We need to know about them to be able to fix them, so
       we're quite dependent on your bug reports! When you do report suspected bugs  in  libcurl,
       please  include  as many details as you possibly can: a protocol dump that CURLOPT_VERBOSE
       produces, library version, as much as possible of your code that uses  libcurl,  operating
       system name and version, compiler name and version etc.

       If  CURLOPT_VERBOSE  is  not enough, you increase the level of debug data your application
       receive by using the CURLOPT_DEBUGFUNCTION.

       Getting some in-depth knowledge about the protocols involved is never wrong, and if you're
       trying to do funny things, you might very well understand libcurl and how to use it better
       if you study the appropriate RFC documents at least briefly.


Upload Data to a Remote Site
       libcurl tries to keep a protocol independent approach to most transfers, thus uploading to
       a remote FTP site is very similar to uploading data to a HTTP server with a PUT request.

       Of course, first you either create an easy handle or you re-use one existing one. Then you
       set the URL to operate on just like before. This is the  remote  URL,  that  we  now  will
       upload.

       Since  we write an application, we most likely want libcurl to get the upload data by ask-
       ing us for it. To make it do that, we set the read callback and the custom pointer libcurl
       will pass to our read callback. The read callback should have a prototype similar to:

        size_t function(char *bufptr, size_t size, size_t nitems, void *userp);

       Where  bufptr is the pointer to a buffer we fill in with data to upload and size*nitems is
       the size of the buffer and therefore also the maximum amount of  data  we  can  return  to
       libcurl  in  this  call.  The  'userp'  pointer is the custom pointer we set to point to a
       struct of ours to pass private data between the application and the callback.

        curl_easy_setopt(easyhandle, CURLOPT_READFUNCTION, read_function);

        curl_easy_setopt(easyhandle, CURLOPT_READDATA, &filedata);

       Tell libcurl that we want to upload:

        curl_easy_setopt(easyhandle, CURLOPT_UPLOAD, 1L);

       A few protocols won't behave properly when uploads are done without any prior knowledge of
       the  expected  file  size. So, set the upload file size using the CURLOPT_INFILESIZE_LARGE
       for all known file sizes like this[1]:

        /* in this example, file_size must be an curl_off_t variable */
        curl_easy_setopt(easyhandle, CURLOPT_INFILESIZE_LARGE, file_size);

       When you call curl_easy_perform(3) this time, it'll perform all the  necessary  operations
       and  when  it  has invoked the upload it'll call your supplied callback to get the data to
       upload. The program should return as much data as possible in every  invoke,  as  that  is
       likely to make the upload perform as fast as possible. The callback should return the num-
       ber of bytes it wrote in the buffer. Returning 0 will signal the end of the upload.


Passwords
       Many protocols use or even require that user name and password are provided to be able  to
       download or upload the data of your choice. libcurl offers several ways to specify them.

       Most  protocols  support that you specify the name and password in the URL itself. libcurl
       will detect this and use them accordingly. This is written like this:

        protocol://user:password AT example.com/path/

       If you need any odd letters in your user name or  password,  you  should  enter  them  URL
       encoded, as %XX where XX is a two-digit hexadecimal number.

       libcurl  also  provides  options  to  set various passwords. The user name and password as
       shown embedded in the URL can instead get set with the CURLOPT_USERPWD option.  The  argu-
       ment  passed to libcurl should be a char * to a string in the format "user:password". In a
       manner like this:

        curl_easy_setopt(easyhandle, CURLOPT_USERPWD, "myname:thesecret");

       Another case where name and password might be needed at times, is for those users who need
       to  authenticate  themselves  to a proxy they use. libcurl offers another option for this,
       the CURLOPT_PROXYUSERPWD. It is used quite similar  to  the  CURLOPT_USERPWD  option  like
       this:

        curl_easy_setopt(easyhandle, CURLOPT_PROXYUSERPWD, "myname:thesecret");

       There's a long time UNIX "standard" way of storing ftp user names and passwords, namely in
       the $HOME/.netrc file. The file should be made private so that only the user may  read  it
       (see  also  the  "Security  Considerations"  chapter), as it might contain the password in
       plain text. libcurl has the ability to use this file to figure out what set of  user  name
       and  password  to  use for a particular host. As an extension to the normal functionality,
       libcurl also supports this file for non-FTP protocols such as HTTP. To make curl use  this
       file, use the CURLOPT_NETRC option:

        curl_easy_setopt(easyhandle, CURLOPT_NETRC, 1L);

       And a very basic example of how such a .netrc file may look like:

        machine myhost.mydomain.com
        login userlogin
        password secretword

       All  these  examples have been cases where the password has been optional, or at least you
       could leave it out and have libcurl attempt to do its job without it. There are times when
       the  password  isn't optional, like when you're using an SSL private key for secure trans-
       fers.

       To pass the known private key password to libcurl:

        curl_easy_setopt(easyhandle, CURLOPT_KEYPASSWD, "keypassword");


HTTP Authentication
       The previous chapter showed how to set user  name  and  password  for  getting  URLs  that
       require  authentication.  When  using  the  HTTP protocol, there are many different ways a
       client can provide those credentials to the server and you can control which  way  libcurl
       will  (attempt  to)  use  them.  The default HTTP authentication method is called 'Basic',
       which is sending the name and password in clear-text in the HTTP request,  base64-encoded.
       This is insecure.

       At  the time of this writing, libcurl can be built to use: Basic, Digest, NTLM, Negotiate,
       GSS-Negotiate and SPNEGO. You can tell libcurl which one to use with  CURLOPT_HTTPAUTH  as
       in:

        curl_easy_setopt(easyhandle, CURLOPT_HTTPAUTH, CURLAUTH_DIGEST);

       And when you send authentication to a proxy, you can also set authentication type the same
       way but instead with CURLOPT_PROXYAUTH:

        curl_easy_setopt(easyhandle, CURLOPT_PROXYAUTH, CURLAUTH_NTLM);

       Both these options allow you to set multiple types  (by  ORing  them  together),  to  make
       libcurl pick the most secure one out of the types the server/proxy claims to support. This
       method does however add a round-trip since libcurl must first ask the server what it  sup-
       ports:

        curl_easy_setopt(easyhandle, CURLOPT_HTTPAUTH,
        CURLAUTH_DIGEST|CURLAUTH_BASIC);

       For  convenience,  you  can use the 'CURLAUTH_ANY' define (instead of a list with specific
       types) which allows libcurl to use whatever method it wants.

       When asking for multiple types, libcurl will pick the available one it considers "best" in
       its own internal order of preference.


HTTP POSTing
       We  get many questions regarding how to issue HTTP POSTs with libcurl the proper way. This
       chapter will thus include examples using both different versions of HTTP POST that libcurl
       supports.

       The  first version is the simple POST, the most common version, that most HTML pages using
       the <form> tag uses. We provide a pointer to the data and tell libcurl to post it  all  to
       the remote site:

           char *data="name=daniel&project=curl";
           curl_easy_setopt(easyhandle, CURLOPT_POSTFIELDS, data);
           curl_easy_setopt(easyhandle, CURLOPT_URL, "http://posthere.com/");

           curl_easy_perform(easyhandle); /* post away! */

       Simple enough, huh? Since you set the POST options with the CURLOPT_POSTFIELDS, this auto-
       matically switches the handle to use POST in the upcoming request.

       Ok, so what if you want to post binary data that also requires you  to  set  the  Content-
       Type:  header  of  the  post?  Well,  binary  posts  prevent libcurl from being able to do
       strlen() on the data to figure out the size, so therefore we must tell libcurl the size of
       the  post data. Setting headers in libcurl requests are done in a generic way, by building
       a list of our own headers and then passing that list to libcurl.

        struct curl_slist *headers=NULL;
        headers = curl_slist_append(headers, "Content-Type: text/xml");

        /* post binary data */
        curl_easy_setopt(easyhandle, CURLOPT_POSTFIELDS, binaryptr);

        /* set the size of the postfields data */
        curl_easy_setopt(easyhandle, CURLOPT_POSTFIELDSIZE, 23L);

        /* pass our list of custom made headers */
        curl_easy_setopt(easyhandle, CURLOPT_HTTPHEADER, headers);

        curl_easy_perform(easyhandle); /* post away! */

        curl_slist_free_all(headers); /* free the header list */

       While the simple examples above cover the majority of all cases where HTTP POST operations
       are  required, they don't do multi-part formposts. Multi-part formposts were introduced as
       a better way to post (possibly large) binary data and were first documented in the RFC1867
       (updated in RFC2388). They're called multi-part because they're built by a chain of parts,
       each part being a single unit of data. Each part has its own name and contents. You can in
       fact create and post a multi-part formpost with the regular libcurl POST support described
       above, but that would require that you build a formpost yourself and provide  to  libcurl.
       To  make that easier, libcurl provides curl_formadd(3). Using this function, you add parts
       to the form. When you're done adding parts, you post the whole form.

       The following example sets two simple text parts with plain textual contents, and  then  a
       file with binary contents and uploads the whole thing.

        struct curl_httppost *post=NULL;
        struct curl_httppost *last=NULL;
        curl_formadd(&post, &last,
                     CURLFORM_COPYNAME, "name",
                     CURLFORM_COPYCONTENTS, "daniel", CURLFORM_END);
        curl_formadd(&post, &last,
                     CURLFORM_COPYNAME, "project",
                     CURLFORM_COPYCONTENTS, "curl", CURLFORM_END);
        curl_formadd(&post, &last,
                     CURLFORM_COPYNAME, "logotype-image",
                     CURLFORM_FILECONTENT, "curl.png", CURLFORM_END);

        /* Set the form info */
        curl_easy_setopt(easyhandle, CURLOPT_HTTPPOST, post);

        curl_easy_perform(easyhandle); /* post away! */

        /* free the post data again */
        curl_formfree(post);

       Multipart  formposts are chains of parts using MIME-style separators and headers. It means
       that each one of these separate parts get a few headers set that describe  the  individual
       content-type,  size etc. To enable your application to handicraft this formpost even more,
       libcurl allows you to supply your own set of custom headers to  such  an  individual  form
       part. You can of course supply headers to as many parts as you like, but this little exam-
       ple will show how you set headers to one specific part when you add that to the post  han-
       dle:

        struct curl_slist *headers=NULL;
        headers = curl_slist_append(headers, "Content-Type: text/xml");

        curl_formadd(&post, &last,
                     CURLFORM_COPYNAME, "logotype-image",
                     CURLFORM_FILECONTENT, "curl.xml",
                     CURLFORM_CONTENTHEADER, headers,
                     CURLFORM_END);

        curl_easy_perform(easyhandle); /* post away! */

        curl_formfree(post); /* free post */
        curl_slist_free_all(headers); /* free custom header list */

       Since  all  options on an easyhandle are "sticky", they remain the same until changed even
       if you do call curl_easy_perform(3), you may need to tell curl to go back to a  plain  GET
       request  if  you intend to do one as your next request. You force an easyhandle to go back
       to GET by using the CURLOPT_HTTPGET option:

        curl_easy_setopt(easyhandle, CURLOPT_HTTPGET, 1L);

       Just setting CURLOPT_POSTFIELDS to "" or NULL will *not* stop libcurl from doing  a  POST.
       It will just make it POST without any data to send!


Showing Progress
       For  historical and traditional reasons, libcurl has a built-in progress meter that can be
       switched on and then makes it present a progress meter in your terminal.

       Switch on the progress meter by, oddly enough, setting CURLOPT_NOPROGRESS  to  zero.  This
       option is set to 1 by default.

       For  most applications however, the built-in progress meter is useless and what instead is
       interesting is the ability to specify a progress callback. The function pointer  you  pass
       to  libcurl  will then be called on irregular intervals with information about the current
       transfer.

       Set the progress callback by using CURLOPT_PROGRESSFUNCTION. And pass a pointer to a func-
       tion that matches this prototype:

        int progress_callback(void *clientp,
                              double dltotal,
                              double dlnow,
                              double ultotal,
                              double ulnow);

       If  any  of  the  input  arguments is unknown, a 0 will be passed. The first argument, the
       'clientp' is the pointer you pass to  libcurl  with  CURLOPT_PROGRESSDATA.  libcurl  won't
       touch it.


libcurl with C++
       There's  basically  only one thing to keep in mind when using C++ instead of C when inter-
       facing libcurl:

       The callbacks CANNOT be non-static class member functions

       Example C++ code:

       class AClass {
           static size_t write_data(void *ptr, size_t size, size_t nmemb,
                                    void *ourpointer)
           {
             /* do what you want with the data */
           }
        }


Proxies
       What "proxy" means according to Merriam-Webster: "a person authorized to act for  another"
       but  also  "the  agency,  function,  or  office  of  a deputy who acts as a substitute for
       another".

       Proxies are exceedingly common these days. Companies often only offer Internet  access  to
       employees  through  their  proxies. Network clients or user-agents ask the proxy for docu-
       ments, the proxy does the actual request and then it returns them.

       libcurl supports SOCKS and HTTP proxies. When a given URL is wanted, libcurl will ask  the
       proxy for it instead of trying to connect to the actual host identified in the URL.

       If  you're using a SOCKS proxy, you may find that libcurl doesn't quite support all opera-
       tions through it.

       For HTTP proxies: the fact that the proxy is a HTTP proxy  puts  certain  restrictions  on
       what  can  actually  happen. A requested URL that might not be a HTTP URL will be still be
       passed to the HTTP proxy to deliver back to libcurl. This happens  transparently,  and  an
       application  may  not  need to know. I say "may", because at times it is very important to
       understand that all operations over a HTTP proxy use the HTTP protocol. For  example,  you
       can't invoke your own custom FTP commands or even proper FTP directory listings.


       Proxy Options

              To tell libcurl to use a proxy at a given port number:

               curl_easy_setopt(easyhandle, CURLOPT_PROXY, "proxy-host.com:8080");

              Some  proxies  require  user authentication before allowing a request, and you pass
              that information similar to this:

               curl_easy_setopt(easyhandle, CURLOPT_PROXYUSERPWD, "user:password");

              If you want to, you can specify the host name only in the CURLOPT_PROXY option, and
              set the port number separately with CURLOPT_PROXYPORT.

              Tell  libcurl  what  kind  of  proxy  it is with CURLOPT_PROXYTYPE (if not, it will
              default to assume a HTTP proxy):

               curl_easy_setopt(easyhandle, CURLOPT_PROXYTYPE, CURLPROXY_SOCKS4);


       Environment Variables

              libcurl automatically checks and uses a set of environment variables to  know  what
              proxies  to  use for certain protocols. The names of the variables are following an
              ancient de facto standard and are built up as "[protocol]_proxy"  (note  the  lower
              casing). Which makes the variable 'http_proxy' checked for a name of a proxy to use
              when the input URL is HTTP. Following the same rule, the variable named 'ftp_proxy'
              is  checked for FTP URLs. Again, the proxies are always HTTP proxies, the different
              names of the variables simply allows different HTTP proxies to be used.

              The  proxy  environment  variable  contents  should  be  in  the  format   "[proto-
              col://][user:password@]machine[:port]".   Where  the  protocol://  part  is  simply
              ignored if present (so http://proxy and bluerk://proxy will do the  same)  and  the
              optional port number specifies on which port the proxy operates on the host. If not
              specified, the internal default port number will be used and that  is  most  likely
              *not* the one you would like it to be.

              There are two special environment variables. 'all_proxy' is what sets proxy for any
              URL in case the protocol specific variable wasn't set,  and  'no_proxy'  defines  a
              list  of  hosts  that  should not use a proxy even though a variable may say so. If
              'no_proxy' is a plain asterisk ("*") it matches all hosts.

              To explicitly disable libcurl's checking for and using the proxy environment  vari-
              ables, set the proxy name to "" - an empty string - with CURLOPT_PROXY.

       SSL and Proxies

              SSL  is  for secure point-to-point connections. This involves strong encryption and
              similar things, which effectively makes it impossible for a proxy to operate  as  a
              "man  in  between" which the proxy's task is, as previously discussed. Instead, the
              only way to have SSL work over a HTTP proxy is to ask the proxy  to  tunnel  trough
              everything without being able to check or fiddle with the traffic.

              Opening  an  SSL  connection  over  a HTTP proxy is therefor a matter of asking the
              proxy for a straight connection to the target host on a  specified  port.  This  is
              made  with  the  HTTP request CONNECT. ("please mr proxy, connect me to that remote
              host").

              Because of the nature of this operation, where the proxy has no idea what  kind  of
              data  that  is  passed in and out through this tunnel, this breaks some of the very
              few advantages that come from using a proxy, such as caching.   Many  organizations
              prevent this kind of tunneling to other destination port numbers than 443 (which is
              the default HTTPS port number).


       Tunneling Through Proxy
              As explained above, tunneling is required for SSL to work and often even restricted
              to the operation intended for SSL; HTTPS.

              This  is  however  not the only time proxy-tunneling might offer benefits to you or
              your application.

              As tunneling opens a direct connection from your application to the remote machine,
              it  suddenly  also  re-introduces the ability to do non-HTTP operations over a HTTP
              proxy. You can in fact use things such as FTP upload or FTP  custom  commands  this
              way.

              Again,  this  is  often  prevented  by  the administrators of proxies and is rarely
              allowed.

              Tell libcurl to use proxy tunneling like this:

               curl_easy_setopt(easyhandle, CURLOPT_HTTPPROXYTUNNEL, 1L);

              In fact, there might even be times when you want to do plain HTTP operations  using
              a  tunnel like this, as it then enables you to operate on the remote server instead
              of asking the proxy to do so. libcurl will not stand in the way for such innovative
              actions either!


       Proxy Auto-Config

              Netscape  first came up with this. It is basically a web page (usually using a .pac
              extension) with a Javascript that when executed by the browser with  the  requested
              URL  as input, returns information to the browser on how to connect to the URL. The
              returned information might be "DIRECT" (which  means  no  proxy  should  be  used),
              "PROXY  host:port" (to tell the browser where the proxy for this particular URL is)
              or "SOCKS host:port" (to direct the browser to a SOCKS proxy).

              libcurl has no means to interpret or evaluate Javascript and thus it  doesn't  sup-
              port  this.  If you get yourself in a position where you face this nasty invention,
              the following advice have been mentioned and used in the past:

              - Depending on the Javascript complexity, write up a script that translates  it  to
              another language and execute that.

              - Read the Javascript code and rewrite the same logic in another language.

              -  Implement  a  Javascript  interpreter; people have successfully used the Mozilla
              Javascript engine in the past.

              - Ask your admins to stop this, for a static proxy setup or similar.


Persistence Is The Way to Happiness
       Re-cycling the same easy handle several times when doing multiple requests is the  way  to
       go.

       After  each  single curl_easy_perform(3) operation, libcurl will keep the connection alive
       and open. A subsequent request using the same easy handle to the same host might  just  be
       able to use the already open connection! This reduces network impact a lot.

       Even  if  the connection is dropped, all connections involving SSL to the same host again,
       will benefit from libcurl's session ID cache that drastically reduces re-connection time.

       FTP connections that are kept alive save a lot of time, as the  command-  response  round-
       trips  are  skipped,  and  also you don't risk getting blocked without permission to login
       again like on many FTP servers only allowing N persons to be logged in at the same time.

       libcurl caches DNS name resolving results, to make lookups of a previously looked up  name
       a lot faster.

       Other  interesting  details  that  improve performance for subsequent requests may also be
       added in the future.

       Each easy handle will attempt to keep the last few connections alive for a while  in  case
       they  are  to be used again. You can set the size of this "cache" with the CURLOPT_MAXCON-
       NECTS option. Default is 5. There is very seldom any point in changing this value, and  if
       you think of changing this it is often just a matter of thinking again.

       To  force  your  upcoming  request to not use an already existing connection (it will even
       close one first if there happens to be one alive to the same host you're about to  operate
       on),  you  can do that by setting CURLOPT_FRESH_CONNECT to 1. In a similar spirit, you can
       also forbid the upcoming request to be "lying" around and possibly get re-used  after  the
       request by setting CURLOPT_FORBID_REUSE to 1.


HTTP Headers Used by libcurl
       When  you  use libcurl to do HTTP requests, it'll pass along a series of headers automati-
       cally. It might be good for you to know and understand these. You can  replace  or  remove
       them by using the CURLOPT_HTTPHEADER option.


       Host   This  header  is  required  by HTTP 1.1 and even many 1.0 servers and should be the
              name of the server we want to talk to. This includes the port  number  if  anything
              but default.


       Accept "*/*".


       Expect When  doing  POST  requests,  libcurl sets this header to "100-continue" to ask the
              server for an "OK" message before it proceeds with sending the  data  part  of  the
              post.  If  the  POSTed  data  amount  is  deemed "small", libcurl will not use this
              header.


Customizing Operations
       There is an ongoing development today where more and more protocols are  built  upon  HTTP
       for transport. This has obvious benefits as HTTP is a tested and reliable protocol that is
       widely deployed and has excellent proxy-support.

       When you use one of these protocols, and even when doing other kinds  of  programming  you
       may  need  to  change the traditional HTTP (or FTP or...)  manners. You may need to change
       words, headers or various data.

       libcurl is your friend here too.


       CUSTOMREQUEST
              If just changing the actual HTTP request keyword is what you want, like  when  GET,
              HEAD or POST is not good enough for you, CURLOPT_CUSTOMREQUEST is there for you. It
              is very simple to use:

               curl_easy_setopt(easyhandle, CURLOPT_CUSTOMREQUEST, "MYOWNREQUEST");

              When using the custom request, you change the request keyword of the actual request
              you are performing. Thus, by default you make a GET request but you can also make a
              POST operation (as described before) and then replace the POST keyword if you  want
              to. You're the boss.


       Modify Headers
              HTTP-like  protocols pass a series of headers to the server when doing the request,
              and you're free to pass any amount of extra headers  that  you  think  fit.  Adding
              headers is this easy:

               struct curl_slist *headers=NULL; /* init to NULL is important */

               headers = curl_slist_append(headers, "Hey-server-hey: how are you?");
               headers = curl_slist_append(headers, "X-silly-content: yes");

               /* pass our list of custom made headers */
               curl_easy_setopt(easyhandle, CURLOPT_HTTPHEADER, headers);

               curl_easy_perform(easyhandle); /* transfer http */

               curl_slist_free_all(headers); /* free the header list */

              ...  and  if you think some of the internally generated headers, such as Accept: or
              Host: don't contain the data you want them to contain, you can replace them by sim-
              ply setting them too:

               headers = curl_slist_append(headers, "Accept: Agent-007");
               headers = curl_slist_append(headers, "Host: munged.host.line");


       Delete Headers
              If  you  replace an existing header with one with no contents, you will prevent the
              header from being sent. For  instance,  if  you  want  to  completely  prevent  the
              "Accept:" header from being sent, you can disable it with code similar to this:

               headers = curl_slist_append(headers, "Accept:");

              Both replacing and canceling internal headers should be done with careful consider-
              ation and you should be aware that you may violate the HTTP protocol when doing so.


       Enforcing chunked transfer-encoding

              By making sure a request uses the custom header "Transfer-Encoding:  chunked"  when
              doing  a non-GET HTTP operation, libcurl will switch over to "chunked" upload, even
              though the size of the data to upload might be known. By default,  libcurl  usually
              switches over to chunked upload automatically if the upload data size is unknown.


       HTTP Version

              All  HTTP  requests includes the version number to tell the server which version we
              support. libcurl speaks HTTP 1.1 by default. Some very old servers don't like  get-
              ting 1.1-requests and when dealing with stubborn old things like that, you can tell
              libcurl to use 1.0 instead by doing something like this:

               curl_easy_setopt(easyhandle, CURLOPT_HTTP_VERSION, CURL_HTTP_VERSION_1_0);


       FTP Custom Commands

              Not all protocols are HTTP-like, and thus the above may not help you when you  want
              to make, for example, your FTP transfers to behave differently.

              Sending  custom  commands  to a FTP server means that you need to send the commands
              exactly as the FTP server expects them (RFC959 is a good guide here), and  you  can
              only  use commands that work on the control-connection alone. All kinds of commands
              that require data interchange and thus need  a  data-connection  must  be  left  to
              libcurl's own judgement. Also be aware that libcurl will do its very best to change
              directory to the target directory before doing  any  transfer,  so  if  you  change
              directory  (with  CWD  or  similar) you might confuse libcurl and then it might not
              attempt to transfer the file in the correct remote directory.

              A little example that deletes a given file before an operation:

               headers = curl_slist_append(headers, "DELE file-to-remove");

               /* pass the list of custom commands to the handle */
               curl_easy_setopt(easyhandle, CURLOPT_QUOTE, headers);

               curl_easy_perform(easyhandle); /* transfer ftp data! */

               curl_slist_free_all(headers); /* free the header list */

              If you would instead want this operation (or chain of operations) to happen _after_
              the  data  transfer  took  place the option to curl_easy_setopt(3) would instead be
              called CURLOPT_POSTQUOTE and used the exact same way.

              The custom FTP command will be issued to the server in  the  same  order  they  are
              added  to  the  list,  and  if  a command gets an error code returned back from the
              server, no more commands will be issued and libcurl will bail  out  with  an  error
              code  (CURLE_QUOTE_ERROR).  Note  that  if  you  use CURLOPT_QUOTE to send commands
              before a transfer, no transfer will actually take place when a  quote  command  has
              failed.

              If  you set the CURLOPT_HEADER to 1, you will tell libcurl to get information about
              the target file and output "headers" about it. The headers will be in "HTTP-style",
              looking like they do in HTTP.

              The option to enable headers or to run custom FTP commands may be useful to combine
              with CURLOPT_NOBODY. If this option is set, no actual file content transfer will be
              performed.


       FTP Custom CUSTOMREQUEST
              If  you  do want to list the contents of a FTP directory using your own defined FTP
              command, CURLOPT_CUSTOMREQUEST will do just that. "NLST" is  the  default  one  for
              listing directories but you're free to pass in your idea of a good alternative.


Cookies Without Chocolate Chips
       In  the  HTTP  sense, a cookie is a name with an associated value. A server sends the name
       and value to the client, and expects it to get sent back on every  subsequent  request  to
       the  server  that  matches  the particular conditions set. The conditions include that the
       domain name and path match and that the cookie hasn't become too old.

       In real-world cases, servers send new cookies to replace existing  ones  to  update  them.
       Server use cookies to "track" users and to keep "sessions".

       Cookies  are sent from server to clients with the header Set-Cookie: and they're sent from
       clients to servers with the Cookie: header.

       To just send whatever cookie you want to a server, you can use  CURLOPT_COOKIE  to  set  a
       cookie string like this:

        curl_easy_setopt(easyhandle, CURLOPT_COOKIE, "name1=var1; name2=var2;");

       In many cases, that is not enough. You might want to dynamically save whatever cookies the
       remote server passes to you, and make sure those cookies  are  then  used  accordingly  on
       later requests.

       One way to do this, is to save all headers you receive in a plain file and when you make a
       request, you tell libcurl to read the previous headers to figure out which cookies to use.
       Set the header file to read cookies from with CURLOPT_COOKIEFILE.

       The  CURLOPT_COOKIEFILE  option  also  automatically enables the cookie parser in libcurl.
       Until the cookie parser is enabled, libcurl will not parse or understand incoming  cookies
       and  they  will  just  be ignored. However, when the parser is enabled the cookies will be
       understood and the cookies will be kept in memory and used properly in subsequent requests
       when  the same handle is used. Many times this is enough, and you may not have to save the
       cookies to disk at all. Note that the file you specify to CURLOPT_COOKIEFILE doesn't  have
       to  exist to enable the parser, so a common way to just enable the parser and not read any
       cookies is to use the name of a file you know doesn't exist.

       If you would rather use existing cookies that you've previously received  with  your  Net-
       scape  or  Mozilla  browsers, you can make libcurl use that cookie file as input. The CUR-
       LOPT_COOKIEFILE is used for that too, as libcurl will automatically find out what kind  of
       file it is and act accordingly.

       Perhaps  the  most advanced cookie operation libcurl offers, is saving the entire internal
       cookie state back into a Netscape/Mozilla formatted cookie file. We call that the  cookie-
       jar.  When  you set a file name with CURLOPT_COOKIEJAR, that file name will be created and
       all received cookies will be stored  in  it  when  curl_easy_cleanup(3)  is  called.  This
       enables cookies to get passed on properly between multiple handles without any information
       getting lost.


FTP Peculiarities We Need
       FTP transfers use a second TCP/IP connection for the data transfer. This is usually a fact
       you  can  forget  and  ignore  but at times this fact will come back to haunt you. libcurl
       offers several different ways to customize how the second connection is being made.

       libcurl can either connect to the server a second time or tell the server to connect  back
       to  it.  The first option is the default and it is also what works best for all the people
       behind firewalls, NATs or IP-masquerading setups.  libcurl then tells the server  to  open
       up  a  new  port  and wait for a second connection. This is by default attempted with EPSV
       first, and if that doesn't work it tries PASV instead. (EPSV is an extension to the origi-
       nal FTP spec and does not exist nor work on all FTP servers.)

       You can prevent libcurl from first trying the EPSV command by setting CURLOPT_FTP_USE_EPSV
       to zero.

       In some cases, you will prefer to have the server connect back to you for the second  con-
       nection.  This might be when the server is perhaps behind a firewall or something and only
       allows connections on a single port. libcurl then  informs  the  remote  server  which  IP
       address  and  port number to connect to.  This is made with the CURLOPT_FTPPORT option. If
       you set it to "-", libcurl will use your system's "default IP address". If you want to use
       a  particular IP, you can set the full IP address, a host name to resolve to an IP address
       or even a local network interface name that libcurl will get the IP address from.

       When doing the "PORT" approach, libcurl will attempt to use the EPRT and the  LPRT  before
       trying  PORT,  as  they work with more protocols. You can disable this behavior by setting
       CURLOPT_FTP_USE_EPRT to zero.


Headers Equal Fun
       Some protocols provide "headers", meta-data separated from the normal data. These  headers
       are by default not included in the normal data stream, but you can make them appear in the
       data stream by setting CURLOPT_HEADER to 1.

       What might be even more useful, is libcurl's ability to separate the headers from the data
       and thus make the callbacks differ. You can for example set a different pointer to pass to
       the ordinary write callback by setting CURLOPT_WRITEHEADER.

       Or, you can set an entirely separate function  to  receive  the  headers,  by  using  CUR-
       LOPT_HEADERFUNCTION.

       The  headers  are  passed  to the callback function one by one, and you can depend on that
       fact. It makes it easier for you to add custom header parsers etc.

       "Headers" for FTP transfers equal all the FTP server responses. They aren't actually  true
       headers, but in this case we pretend they are! ;-)


Post Transfer Information
        [ curl_easy_getinfo ]


Security Considerations
       The  libcurl  project  takes  security seriously.  The library is written with caution and
       precautions are taken to mitigate many kinds of risks  encountered  while  operating  with
       potentially  malicious  servers on the Internet.  It is a powerful library, however, which
       allows application writers to make trade offs between ease  of  writing  and  exposure  to
       potential  risky  operations.  If used the right way, you can use libcurl to transfer data
       pretty safely.

       Many applications are used in closed networks where users and servers can be trusted,  but
       many  others  are  used  on arbitrary servers and are fed input from potentially untrusted
       users.  Following is a discussion about some risks in the ways in which applications  com-
       monly  use  libcurl and potential mitigations of those risks. It is by no means comprehen-
       sive, but shows classes of attacks that robust applications should  consider.  The  Common
       Weakness  Enumeration  project  at  http://cwe.mitre.org/  is a good reference for many of
       these and similar types of weaknesses of which application writers should be aware.


       Command Lines
              If you use a command line tool (such as curl)  that  uses  libcurl,  and  you  give
              options  to  the tool on the command line those options can very likely get read by
              other users of your system when they use 'ps' or other tools to list currently run-
              ning processes.

              To  avoid  this problem, never feed sensitive things to programs using command line
              options. Write them to a protected file and use the -K option to avoid this.


       .netrc .netrc is a pretty handy file/feature that allows you to login quickly and automat-
              ically  to  frequently visited sites. The file contains passwords in clear text and
              is a real security risk. In some cases, your .netrc is also stored in a home direc-
              tory that is NFS mounted or used on another network based file system, so the clear
              text password will fly through your network every time anyone reads that file!

              To avoid this problem, don't use .netrc files and never store  passwords  in  plain
              text anywhere.


       Clear Text Passwords
              Many  of the protocols libcurl supports send name and password unencrypted as clear
              text (HTTP Basic authentication, FTP, TELNET etc). It is very easy  for  anyone  on
              your  network or a network nearby yours to just fire up a network analyzer tool and
              eavesdrop on your passwords. Don't let the fact that HTTP Basic uses base64 encoded
              passwords  fool  you.  They  may not look readable at a first glance, but they very
              easily "deciphered" by anyone within seconds.

              To avoid this problem, use HTTP authentication  methods  or  other  protocols  that
              don't let snoopers see your password: HTTP with Digest, NTLM or GSS authentication,
              HTTPS, FTPS, SCP, SFTP and FTP-Kerberos are a few examples.


       Redirects
              The CURLOPT_FOLLOWLOCATION option automatically follows HTTP redirects  sent  by  a
              remote server.  These redirects can refer to any kind of URL, not just HTTP.  A re-
              direct to a file: URL would cause the libcurl to read (or  write)  arbitrary  files
              from  the  local  filesystem.  If the application returns the data back to the user
              (as would happen in some kinds of CGI scripts), an attacker could leverage this  to
              read otherwise forbidden data (e.g.  file://localhost/etc/passwd).

              If  authentication  credentials  are stored in the ~/.netrc file, or Kerberos is in
              use, any other URL type (not just file:) that requires authentication  is  also  at
              risk.  A redirect such as ftp://some-internal-server/private-file would then return
              data even when the server is password protected.

              In the same way, if an unencrypted SSH private key has been configured for the user
              running  the  libcurl application, SCP: or SFTP: URLs could access password or pri-
              vate-key protected resources, e.g. sftp://user@some-internal-server/etc/passwd

              The CURLOPT_REDIR_PROTOCOLS and CURLOPT_NETRC  options  can  be  used  to  mitigate
              against this kind of attack.

              A  redirect  can  also  specify  a  location  available only on the machine running
              libcurl, including servers hidden  behind  a  firewall  from  the  attacker.   e.g.
              http://127.0.0.1/  or  http://intranet/delete-stuff.cgi?delete=all or tftp://bootp-
              server/pc-config-data

              Apps can mitigate against this by  disabling  CURLOPT_FOLLOWLOCATION  and  handling
              redirects  itself,  sanitizing  URLs  as necessary. Alternately, an app could leave
              CURLOPT_FOLLOWLOCATION enabled but set CURLOPT_REDIR_PROTOCOLS and install  a  CUR-
              LOPT_OPENSOCKETFUNCTION  callback  function in which addresses are sanitized before
              use.


       Private Resources
              A user who can control the DNS server of a domain being passed in within a URL  can
              change the address of the host to a local, private address which the libcurl appli-
              cation will then use. e.g. The innocuous URL http://fuzzybunnies.example.com/ could
              actually resolve to the IP address of a server behind a firewall, such as 127.0.0.1
              or 10.1.2.3 Apps can mitigate against this by setting a  CURLOPT_OPENSOCKETFUNCTION
              and checking the address before a connection.

              All  the  malicious  scenarios regarding redirected URLs apply just as well to non-
              redirected URLs, if the user is allowed to specify  an  arbitrary  URL  that  could
              point to a private resource. For example, a web app providing a translation service
              might happily translate file://localhost/etc/passwd and display the  result.   Apps
              can  mitigate  against this with the CURLOPT_PROTOCOLS option as well as by similar
              mitigation techniques for redirections.

              A malicious FTP server could in response to the PASV command return an  IP  address
              and  port  number  for a server local to the app running libcurl but behind a fire-
              wall.  Apps can mitigate against this by using the CURLOPT_FTP_SKIP_PASV_IP  option
              or CURLOPT_FTPPORT.


       Uploads
              When  uploading,  a  redirect can cause a local (or remote) file to be overwritten.
              Apps must not allow any unsanitized URL to be passed in for  uploads.   Also,  CUR-
              LOPT_FOLLOWLOCATION  should not be used on uploads.  Instead, the app should handle
              redirects itself, sanitizing each URL first.


       Authentication
              Use of CURLOPT_UNRESTRICTED_AUTH could cause authentication information to be  sent
              to  an  unknown  second  server.   Apps can mitigate against this by disabling CUR-
              LOPT_FOLLOWLOCATION and handling redirects itself, sanitizing where necessary.

              Use of the CURLAUTH_ANY option to CURLOPT_HTTPAUTH could result in  user  name  and
              password being sent in clear text to an HTTP server.  Instead, use CURLAUTH_ANYSAFE
              which ensures that the password is encrypted over the network,  or  else  fail  the
              request.

              Use  of  the CURLUSESSL_TRY option to CURLOPT_USE_SSL could result in user name and
              password being sent in clear text to an FTP server.  Instead,  use  CURLUSESSL_CON-
              TROL to ensure that an encrypted connection is used or else fail the request.


       Cookies
              If  cookies  are  enabled  and cached, then a user could craft a URL which performs
              some malicious action to a site whose authentication is already stored in a cookie.
              e.g.  http://mail.example.com/delete-stuff.cgi?delete=all Apps can mitigate against
              this by disabling cookies or clearing them between requests.


       Dangerous URLs
              SCP URLs can contain raw commands within the scp: URL, which is a  side  effect  of
              how  the  SCP  protocol  is designed. e.g.  scp://user:pass@host/a;date >/tmp/test;
              Apps must not allow unsanitized SCP: URLs to be passed in for downloads.


       Denial of Service
              A malicious server could cause libcurl to effectively hang by sending a trickle  of
              data  through,  or  even  no  data at all but just keeping the TCP connection open.
              This could result in a denial-of-service attack. The  CURLOPT_TIMEOUT  and/or  CUR-
              LOPT_LOW_SPEED_LIMIT options can be used to mitigate against this.

              A  malicious  server  could  cause  libcurl to effectively hang by starting to send
              data, then severing the connection without cleanly closing the TCP connection.  The
              app  could  install  a  CURLOPT_SOCKOPTFUNCTION  callback  function and set the TCP
              SO_KEEPALIVE option to mitigate against this.  Setting one of the  timeout  options
              would also work against this attack.

              A  malicious  server  could  cause  libcurl to download an infinite amount of data,
              potentially causing all of memory or disk to be filled.  Setting  the  CURLOPT_MAX-
              FILESIZE_LARGE  option  is  not sufficient to guard against this.  Instead, the app
              should monitor the amount of data received within the write  or  progress  callback
              and abort once the limit is reached.

              A malicious HTTP server could cause an infinite redirection loop, causing a denial-
              of-service. This can be mitigated by using the CURLOPT_MAXREDIRS option.


       Arbitrary Headers
              User-supplied data must be sanitized when used in options  like  CURLOPT_USERAGENT,
              CURLOPT_HTTPHEADER,  CURLOPT_POSTFIELDS and others that are used to generate struc-
              tured data. Characters like embedded carriage returns or ampersands could allow the
              user  to  create  additional  headers or fields that could cause malicious transac-
              tions.


       Server-supplied Names
              A server can supply data which the application may, in some cases, use  as  a  file
              name.  The  curl  command-line  tool does this with --remote-header-name, using the
              Content-disposition: header to generate a file name.  An application could also use
              CURLINFO_EFFECTIVE_URL to generate a file name from a server-supplied redirect URL.
              Special care must be taken to sanitize such names to avoid  the  possibility  of  a
              malicious server supplying one like "/etc/passwd", "utoexec.bat" or even ".bashrc".


       Server Certificates
              A  secure application should never use the CURLOPT_SSL_VERIFYPEER option to disable
              certificate validation. There are numerous attacks that are enabled  by  apps  that
              fail  to  properly  validate server TLS/SSL certificates, thus enabling a malicious
              server to spoof a legitimate one. HTTPS without validated  certificates  is  poten-
              tially as insecure as a plain HTTP connection.


       Showing What You Do
              On  a  related  issue, be aware that even in situations like when you have problems
              with libcurl and ask someone for help, everything you reveal in order to  get  best
              possible  help  might  also impose certain security related risks. Host names, user
              names, paths, operating system specifics, etc (not to mention passwords of  course)
              may in fact be used by intruders to gain additional information of a potential tar-
              get.

              To avoid this problem, you must of course use your common  sense.  Often,  you  can
              just  edit out the sensitive data or just search/replace your true information with
              faked data.


Multiple Transfers Using the multi Interface
       The easy interface as described in detail in this document is a synchronous interface that
       transfers one file at a time and doesn't return until it is done.

       The  multi interface, on the other hand, allows your program to transfer multiple files in
       both directions at the same time, without forcing you to use multiple threads.   The  name
       might  make it seem that the multi interface is for multi-threaded programs, but the truth
       is almost the reverse.  The multi interface can allow  a  single-threaded  application  to
       perform  the  same  kinds of multiple, simultaneous transfers that multi-threaded programs
       can perform.  It allows many of the benefits of multi-threaded transfers without the  com-
       plexity of managing and synchronizing many threads.

       To use this interface, you are better off if you first understand the basics of how to use
       the easy interface. The multi interface is simply a way to make multiple transfers at  the
       same time by adding up multiple easy handles into a "multi stack".

       You  create  the easy handles you want and you set all the options just like you have been
       told above, and then you create a multi handle with curl_multi_init(3) and add  all  those
       easy handles to that multi handle with curl_multi_add_handle(3).

       When  you've  added the handles you have for the moment (you can still add new ones at any
       time), you start the transfers by calling curl_multi_perform(3).

       curl_multi_perform(3) is asynchronous. It will only execute as little as possible and then
       return back control to your program. It is designed to never block.

       The  best  usage of this interface is when you do a select() on all possible file descrip-
       tors or sockets to know when to call libcurl again. This also makes it  easy  for  you  to
       wait and respond to actions on your own application's sockets/handles. You figure out what
       to select() for by using curl_multi_fdset(3), that fills in a set of fd_set variables  for
       you with the particular file descriptors libcurl uses for the moment.

       When  you  then call select(), it'll return when one of the file handles signal action and
       you then call curl_multi_perform(3) to allow libcurl to do what it wants to do. Take  note
       that  libcurl does also feature some time-out code so we advise you to never use very long
       timeouts on select() before you call curl_multi_perform(3), which thus  should  be  called
       unconditionally  every  now  and  then  even if none of its file descriptors have signaled
       ready. Another precaution you should  use:  always  call  curl_multi_fdset(3)  immediately
       before the select() call since the current set of file descriptors may change when calling
       a curl function.

       If you want to stop the transfer of one of the easy handles in  the  stack,  you  can  use
       curl_multi_remove_handle(3)  to remove individual easy handles. Remember that easy handles
       should be curl_easy_cleanup(3)ed.

       When a transfer within the multi stack has finished, the counter of running transfers  (as
       filled  in  by  curl_multi_perform(3))  will  decrease.  When the number reaches zero, all
       transfers are done.

       curl_multi_info_read(3) can be used to get information about completed transfers. It  then
       returns  the  CURLcode  for each easy transfer, to allow you to figure out success on each
       individual transfer.


SSL, Certificates and Other Tricks
        [ seeding, passwords, keys, certificates, ENGINE, ca certs ]


Sharing Data Between Easy Handles
       You can share some data between easy handles when the easy interface  is  used,  and  some
       data is share automatically when you use the multi interface.

       When you add easy handles to a multi handle, these easy handles will automatically share a
       lot of the data that otherwise would be kept on a per-easy  handle  basis  when  the  easy
       interface is used.

       The  DNS  cache  is  shared  between handles within a multi handle, making subsequent name
       resolvings faster and the connection pool that is kept to better allow persistent  connec-
       tions  and  connection re-use is shared. If you're using the easy interface, you can still
       share these between specific easy handles by  using  the  share  interface,  see  libcurl-
       share(3).

       Some  things  are  never  shared automatically, not within multi handles, like for example
       cookies so the only way to share that is with the share interface.

Footnotes
       [1]    libcurl 7.10.3 and later have the ability to switch over to chunked Transfer-Encod-
              ing in cases where HTTP uploads are done with data of an unknown size.

       [2]    This  happens on Windows machines when libcurl is built and used as a DLL. However,
              you can still do this on Windows if you link with a static library.

       [3]    The curl-config tool is generated at build-time (on UNIX-like systems)  and  should
              be  installed  with  the  'make  install'  or similar instruction that installs the
              library, header files, man pages etc.

       [4]    This behavior was different in versions before 7.17.0, where strings had to  remain
              valid past the end of the curl_easy_setopt(3) call.



libcurl                                     4 Mar 2009                        libcurl-tutorial(3)

Generated by $Id: phpMan.php,v 4.55 2007/09/05 04:42:51 chedong Exp $ Author: Che Dong
On Apache
Under GNU General Public License
2024-05-01 16:00 @18.191.102.112 CrawledBy Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ClaudeBot/1.0; +claudebot@anthropic.com)
Valid XHTML 1.0!Valid CSS!