# Statements ## Prepared statement Consider the following examples: ```cpp // Example 1. for (int i = 0; i != 100; ++i) { sql << "insert into numbers(value) values(" << i << ")"; } // Example 2. for (int i = 0; i != 100; ++i) { sql << "insert into numbers(value) values(:val)", use(i); } ``` Both examples will populate the table `numbers` with the values from `0` to `99`. The problem is that in both examples, not only the statement execution is repeated 100 times, but also the statement parsing and preparation. This means unnecessary overhead, even if some of the database servers are likely to optimize the second case. In fact, more complicated queries are likely to suffer in terms of lower performance, because finding the optimal execution plan is quite expensive and here it would be needlessly repeated. ### Statement preparation The following example uses the class `statement` explicitly, by preparing the statement only once and repeating its execution with changing data (note the use of `prepare` member of `session` class): ```cpp int i; statement st = (sql.prepare << "insert into numbers(value) values(:val)", use(i)); for (i = 0; i != 100; ++i) { st.execute(true); } ``` The `true` parameter given to the `execute` method indicates that the actual data exchange is wanted, so that the meaning of the whole example is > "prepare the statement and exchange the data for each value of variable `i`". ### Portability note: The above syntax is supported for all backends, even if some database server does not actually provide this functionality - in which case the library will internally execute the query in a single phase, without really separating the statement preparation from execution. For PostgreSQL servers older than 8.0 it is necessary to define the `SOCI_POSTGRESQL_NOPREPARE` macro while compiling the library to fall back to this one-phase behaviour. Simply, pass `-DSOCI_POSTGRESQL_NOPREPARE=ON` variable to CMake. ## Rowset and iterator The `rowset` class provides an alternative means of executing queries and accessing results using STL-like iterator interface. The `rowset_iterator` type is compatible with requirements defined for input iterator category and is available via `iterator` and `const_iterator` definitions in the `rowset` class. The `rowset` itself can be used only with select queries. The following example creates an instance of the `rowset` class and binds query results into elements of `int` type - in this query only one result column is expected. After executing the query the code iterates through the query result using `rowset_iterator`: ```cpp rowset rs = (sql.prepare << "select values from numbers"); for (rowset::const_iterator it = rs.begin(); it != rs.end(); ++it) { cout << *it << '\n'; } ``` Another example shows how to retrieve more complex results, where `rowset` elements are of type `row` and therefore use [dynamic bindings](exchange.html#dynamic): ```cpp // person table has 4 columns rowset rs = (sql.prepare << "select id, firstname, lastname, gender from person"); // iteration through the resultset: for (rowset::const_iterator it = rs.begin(); it != rs.end(); ++it) { row const& row = *it; // dynamic data extraction from each row: cout << "Id: " << row.get(0) << '\n' << "Name: " << row.get(1) << " " << row.get(2) << '\n' << "Gender: " << row.get(3) << endl; } ``` The `rowset_iterator` can be used with standard algorithms as well: ```cpp rowset rs = (sql.prepare << "select firstname from person"); std::copy(rs.begin(), rs.end(), std::ostream_iterator(std::cout, "\n")); ``` Above, the query result contains a single column which is bound to `rowset` element of type of `std::string`. All records are sent to standard output using the `std::copy` algorithm. ## Bulk operations When using some databases, further performance improvements may be possible by having the underlying database API group operations together to reduce network roundtrips. SOCI makes such bulk operations possible by supporting `std::vector` based types. The following example presents how to insert 100 records in 4 batches. It is also important to note, that size of vector remains equal in every batch interaction. This ensures vector is not reallocated and, what's crucial for the bulk trick, new data should be pushed to the vector before every call to `statement::execute`: ```cpp // Example 3. void fill_ids(std::vector& ids) { for (std::size_t i = 0; i < ids.size(); ++i) ids[i] = i; // mimics source of a new ID } const int BATCH_SIZE = 25; std::vector ids(BATCH_SIZE); statement st = (sql.prepare << "insert into numbers(value) values(:val)", use(ids)); for (int i = 0; i != 4; ++i) { fill_ids(ids); st.execute(true); } ``` Given batch size is 25, this example should insert 4 x 25 = 100 records. (Of course, the size of the vector that will achieve optimum performance will vary, depending on many environmental factors, such as network speed.) It is also possible to read all the numbers written in the above examples: ```cpp int i; statement st = (sql.prepare << "select value from numbers order by value", into(i)); st.execute(); while (st.fetch()) { cout << i << '\n'; } ``` In the above example, the `execute` method is called with the default parameter `false`. This means that the statement should be executed, but the actual data exchange will be performed later. Further `fetch` calls perform the actual data retrieval and cursor traversal. The *end-of-cursor* condition is indicated by the `fetch` function returning `false`. The above code example should be treated as an idiomatic way of reading many rows of data, *one at a time*. It is further possible to select records in batches into `std::vector` based types, with the size of the vector specifying the number of records to retrieve in each round trip: ```cpp std::vector valsOut(100); sql << "select val from numbers", into(valsOut); ``` Above, the value `100` indicates that no more values should be retrieved, even if it would be otherwise possible. If there are less rows than asked for, the vector will be appropriately down-sized. The `statement::execute()` and `statement::fetch()` functions can also be used to repeatedly select all rows returned by a query into a vector based type: ```cpp const int BATCH_SIZE = 30; std::vector valsOut(BATCH_SIZE); statement st = (sql.prepare << "select value from numbers", into(valsOut)); st.execute(); while (st.fetch()) { std::vector::iterator pos; for(pos = valsOut.begin(); pos != valsOut.end(); ++pos) { cout << *pos << '\n'; } valsOut.resize(BATCH_SIZE); } ``` Assuming there are 100 rows returned by the query, the above code will retrieve and print all of them. Since the output vector was created with size 30, it will take (at least) 4 calls to `fetch()` to retrieve all 100 values. Each call to `fetch()` can potentially resize the vector to a size less than its initial size - how often this happens depends on the underlying database implementation. This explains why the `resize(BATCH_SIZE)` operation is needed - it is there to ensure that each time the `fetch()` is called, the vector is ready to accept the next bunch of values. Without this operation, the vector *might* be getting smaller with subsequent iterations of the loop, forcing more iterations to be performed (because *all* rows will be read anyway), than really needed. Note the following details about the above examples: * After performing `fetch()`, the vector's size might be *less* than requested, but `fetch()` returning true means that there was *at least one* row retrieved. * It is forbidden to manually resize the vector to the size *higher* than it was initially (this can cause the vector to reallocate its internal buffer and the library can lose track of it). Taking these points under consideration, the above code example should be treated as an idiomatic way of reading many rows by bunches of requested size. ### Portability note Actually, all supported backends guarantee that the requested number of rows will be read with each fetch and that the vector will never be down-sized, unless for the last fetch, when the end of rowset condition is met. This means that the manual vector resizing is in practice not needed - the vector will keep its size until the end of rowset. The above idiom, however, is provided with future backends in mind, where the constant size of the vector might be too expensive to guarantee and where allowing `fetch` to down-size the vector even before reaching the end of rowset might buy some performance gains. ## Statement caching Some backends have some facilities to improve statement parsing and compilation to limit overhead when creating commonly used query. But for backends that does not support this kind optimization you can keep prepared statement and use it later with new references. To do such, prepare a statement as usual, you have to use `exchange` to bind new variables to statement object, then `execute` statement and finish by cleaning bound references with `bind_clean_up`. ```cpp sql << "CREATE TABLE test(a INTEGER)"; { // prepare statement soci::statement stmt = (db.prepare << "INSERT INTO numbers(value) VALUES(:val)"); { // first insert int a0 = 0; // update reference stmt.exchange(soci::use(a0)); stmt.define_and_bind(); stmt.execute(true); stmt.bind_clean_up(); } { // come later, second insert int a1 = 1; // update reference stmt.exchange(soci::use(a1)); stmt.define_and_bind(); stmt.execute(true); stmt.bind_clean_up(); } } { std::vector v(10); db << "SELECT value FROM numbers", soci::into(v); for (int i = 0; i < v.size(); ++i) std::cout << "value " << i << ": " << v[i] << std::endl; } ```