1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379 2380 2381 2382 2383 2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435 2436 2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506 2507 2508 2509 2510 2511 2512 2513 2514 2515 2516 2517 2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555 2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566 2567 2568 2569 2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593 2594 2595 2596 2597 2598 2599 2600 2601 2602 2603 2604 2605 2606 2607 2608 2609 2610 2611 2612 2613 2614 2615 2616 2617 2618 2619 2620 2621 2622 2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634 2635 2636 2637 2638 2639 2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659 2660 2661 2662 2663 2664 2665 2666 2667 2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680 2681 2682 2683 2684 2685 2686 2687 2688 2689 2690 2691 2692 2693 2694 2695 2696 2697 2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720 2721 2722 2723 2724 2725 2726 2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748 2749 2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777 2778 2779 2780 2781 2782 2783 2784 2785 2786 2787 2788 2789 2790 2791 2792 2793 2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810 2811 2812 2813 2814 2815 2816 2817 2818 2819 2820 2821 2822 2823 2824 2825 2826 2827 2828 2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843 2844 2845 2846 2847 2848 2849 2850 2851 2852 2853 2854 2855 2856 2857 2858 2859 2860 2861 2862 2863 2864 2865 2866 2867 2868 2869 2870 2871 2872 2873 2874 2875 2876 2877 2878 2879 2880 2881 2882 2883 2884 2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901 2902 2903 2904 2905 2906 2907 2908 2909 2910 2911 2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923 2924 2925 2926 2927 2928 2929 2930 2931 2932 2933 2934 2935 2936 2937 2938 2939 2940 2941 2942 2943 2944 2945 2946 2947 2948 2949 2950 2951 2952 2953 2954 2955 2956 2957 2958 2959 2960 2961 2962 2963 2964 2965 2966 2967 2968 2969 2970 2971 2972 2973 2974 2975 2976 2977 2978 2979 2980 2981 2982 2983 2984 2985 2986 2987 2988 2989 2990 2991 2992 2993 2994 2995 2996 2997 2998 2999 3000 3001 3002 3003 3004 3005 3006 3007 3008 3009 3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020 3021 3022 3023 3024 3025 3026 3027 3028 3029 3030 3031 3032 3033 3034 3035 3036 3037 3038 3039 3040 3041 3042 3043 3044 3045 3046 3047 3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078 3079 3080 3081 3082 3083 3084 3085 3086 3087 3088 3089 3090 3091 3092 3093 3094 3095 3096 3097 3098 3099 3100 3101 3102 3103 3104 3105 3106 3107 3108 3109 3110 3111 3112 3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131 3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159 3160 3161 3162 3163 3164 3165 3166 3167 3168 3169 3170 3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206 3207 3208 3209 3210 3211 3212 3213 3214 3215 3216 3217 3218 3219 3220 3221 3222 3223 3224 3225 3226 3227 3228 3229 3230 3231 3232 3233 3234 3235 3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246 3247 3248 3249 3250 3251 3252 3253 3254 3255 3256 3257 3258 3259 3260 3261 3262 3263 3264 3265 3266 3267 3268 3269 3270 3271 3272 3273 3274 3275 3276 3277 3278 3279 3280 3281 3282 3283 3284 3285 3286 3287 3288 3289 3290 3291 3292 3293 3294 3295 3296 3297 3298 3299 3300 3301 3302 3303 3304 3305 3306 3307 3308 3309 3310 3311 3312 3313 3314 3315 3316 3317 3318 3319 3320 3321 3322 3323 3324 3325 3326 3327 3328 3329 3330 3331 3332 3333 3334 3335 3336 3337 3338 3339 3340 3341 3342 3343 3344 3345 3346 3347 3348 3349 3350 3351 3352 3353 3354 3355 3356 3357 3358 3359 3360 3361 3362 3363 3364 3365 3366 3367 3368 3369 3370 3371 3372 3373 3374 3375 3376 3377 3378 3379 3380 3381 3382 3383 3384 3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396 3397 3398 3399 3400 3401 3402 3403 3404 3405 3406 3407 3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419 3420 3421 3422 3423 3424 3425 3426 3427 3428 3429 3430 3431 3432 3433 3434 3435 3436 3437 3438 3439 3440 3441 3442 3443 3444 3445 3446 3447 3448 3449 3450 3451 3452 3453 3454 3455 3456 3457 3458 3459 3460 3461 3462 3463 3464 3465 3466 3467 3468 3469 3470 3471 3472 3473 3474 3475 3476 3477 3478 3479 3480 3481 3482 3483 3484 3485 3486 3487 3488 3489 3490 3491 3492 3493 3494 3495 3496 3497 3498 3499 3500 3501 3502 3503 3504 3505 3506 3507 3508 3509 3510 3511 3512 3513 3514 3515 3516 3517 3518 3519 3520 3521 3522 3523 3524 3525 3526 3527 3528 3529 3530 3531 3532 3533 3534 3535 3536 3537 3538 3539 3540 3541 3542 3543 3544 3545 3546 3547 3548 3549 3550 3551 3552 3553 3554 3555 3556 3557 3558 3559 3560 3561 3562 3563 3564 3565 3566 3567 3568 3569 3570 3571 3572 3573 3574 3575 3576 3577 3578 3579 3580 3581 3582 3583 3584 3585 3586 3587 3588 3589 3590 3591 3592 3593 3594 3595 3596 3597 3598 3599 3600 3601 3602 3603 3604 3605 3606 3607 3608 3609 3610 3611 3612 3613 3614 3615 3616 3617 3618 3619 3620 3621 3622 3623 3624 3625 3626 3627 3628 3629 3630 3631 3632 3633 3634 3635 3636 3637 3638 3639 3640 3641 3642 3643 3644 3645 3646 3647 3648 3649 3650 3651 3652 3653 3654 3655 3656 3657 3658 3659 3660 3661 3662 3663 3664 3665 3666 3667 3668 3669 3670 3671 3672 3673 3674 3675 3676 3677 3678 3679 3680 3681 3682 3683 3684 3685 3686 3687 3688 3689 3690 3691 3692 3693 3694 3695 3696 3697 3698 3699 3700 3701 3702 3703 3704 3705 3706 3707 3708 3709 3710 3711 3712 3713 3714 3715 3716 3717 3718 3719 3720 3721 3722 3723 3724 3725 3726 3727 3728 3729 3730 3731 3732 3733 3734 3735 3736 3737 3738 3739 3740 3741 3742 3743 3744 3745 3746 3747 3748 3749 3750 3751 3752 3753 3754 3755 3756 3757 3758 3759 3760 3761 3762 3763 3764 3765 3766 3767 3768 3769 3770 3771 3772 3773 3774 3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785 3786 3787 3788 3789 3790 3791 3792 3793 3794 3795 3796 3797 3798 3799 3800 3801 3802 3803 3804 3805 3806 3807 3808 3809 3810 3811 3812 3813 3814 3815 3816 3817 3818 3819 3820 3821 3822 3823 3824 3825 3826 3827 3828 3829 3830 3831 3832 3833 3834 3835 3836 3837 3838 3839 3840 3841 3842 3843 3844 3845 3846 3847 3848 3849 3850 3851 3852 3853 3854 3855 3856 3857 3858 3859 3860 3861 3862 3863 3864 3865 3866 3867 3868 3869 3870 3871 3872 3873 3874 3875 3876 3877 3878 3879 3880 3881 3882 3883 3884 3885 3886 3887 3888 3889 3890 3891 3892 3893 3894 3895 3896 3897 3898 3899 3900 3901 3902 3903 3904 3905 3906 3907 3908 3909 3910 3911 3912 3913 3914 3915 3916 3917 3918 3919 3920 3921 3922 3923 3924 3925 3926 3927 3928 3929 3930 3931 3932 3933 3934 3935 3936 3937 3938 3939 3940 3941 3942 3943 3944 3945 3946 3947 3948 3949 3950 3951 3952 3953 3954 3955 3956 3957 3958 3959 3960 3961 3962 3963 3964 3965 3966 3967 3968 3969 3970 3971 3972 3973 3974 3975 3976 3977 3978 3979 3980 3981 3982 3983 3984 3985 3986 3987 3988 3989 3990 3991 3992 3993 3994 3995 3996 3997 3998 3999 4000 4001 4002 4003 4004 4005 4006 4007 4008 4009 4010 4011 4012 4013 4014 4015 4016 4017 4018 4019 4020 4021 4022 4023 4024 4025 4026 4027 4028 4029 4030 4031 4032 4033 4034 4035 4036 4037 4038 4039 4040 4041 4042 4043 4044 4045 4046 4047 4048 4049 4050 4051 4052 4053 4054 4055 4056 4057 4058 4059 4060 4061 4062 4063 4064 4065 4066 4067 4068 4069 4070 4071 4072 4073 4074 4075 4076 4077 4078 4079 4080 4081 4082 4083 4084 4085 4086 4087 4088 4089 4090 4091 4092 4093 4094 4095 4096 4097 4098 4099 4100 4101 4102 4103 4104 4105 4106 4107 4108 4109 4110 4111 4112 4113 4114 4115 4116 4117 4118 4119 4120 4121 4122 4123 4124 4125 4126 4127 4128 4129 4130 4131 4132 4133 4134 4135 4136 4137 4138 4139 4140 4141 4142 4143 4144 4145 4146 4147 4148 4149 4150 4151 4152 4153 4154 4155 4156 4157 4158 4159 4160 4161 4162 4163 4164 4165 4166 4167 4168 4169 4170 4171 4172 4173 4174 4175 4176 4177 4178 4179 4180 4181 4182 4183 4184 4185 4186 4187 4188 4189 4190 4191 4192 4193 4194 4195 4196 4197 4198 4199 4200 4201 4202 4203 4204 4205 4206 4207 4208 4209 4210 4211 4212 4213 4214 4215 4216 4217 4218 4219 4220 4221 4222 4223 4224 4225 4226 4227 4228 4229 4230 4231 4232 4233 4234 4235 4236 4237 4238 4239 4240 4241 4242 4243 4244 4245 4246 4247 4248 4249 4250 4251 4252 4253 4254 4255 4256 4257 4258 4259 4260 4261 4262 4263 4264 4265 4266 4267 4268 4269 4270 4271 4272 4273 4274 4275 4276 4277 4278 4279 4280 4281 4282 4283 4284 4285 4286 4287 4288 4289 4290 4291 4292 4293 4294 4295 4296 4297 4298 4299 4300 4301 4302 4303 4304 4305 4306 4307 4308 4309 4310 4311 4312 4313 4314 4315 4316 4317 4318 4319 4320 4321 4322 4323 4324 4325 4326 4327 4328 4329 4330 4331 4332 4333 4334 4335 4336 4337 4338 4339 4340 4341 4342 4343 4344 4345 4346 4347 4348 4349 4350 4351 4352 4353 4354 4355 4356 4357 4358 4359 4360 4361 4362 4363 4364 4365 4366 4367 4368 4369 4370 4371 4372 4373 4374 4375 4376 4377 4378 4379 4380 4381 4382 4383 4384 4385 4386 4387 4388 4389 4390 4391 4392 4393 4394 4395 4396 4397 4398 4399 4400 4401 4402 4403 4404 4405 4406 4407 4408 4409 4410 4411 4412 4413 4414 4415 4416 4417 4418 4419 4420 4421 4422 4423 4424 4425 4426 4427 4428 4429 4430 4431 4432 4433 4434 4435 4436 4437 4438 4439 4440 4441 4442 4443 4444 4445 4446 4447 4448 4449 4450 4451 4452 4453 4454 4455 4456 4457 4458 4459 4460 4461 4462 4463 4464 4465 4466 4467 4468 4469 4470 4471 4472 4473 4474 4475 4476 4477 4478 4479 4480 4481 4482 4483 4484 4485 4486 4487 4488 4489 4490 4491 4492 4493 4494 4495 4496 4497 4498 4499 4500 4501 4502 4503 4504 4505 4506 4507 4508 4509 4510 4511 4512 4513 4514 4515 4516 4517 4518 4519 4520 4521 4522 4523 4524 4525 4526 4527 4528 4529 4530 4531 4532 4533 4534 4535 4536 4537 4538 4539 4540 4541 4542 4543 4544 4545 4546 4547 4548 4549 4550 4551 4552 4553 4554 4555 4556 4557 4558 4559 4560 4561 4562 4563 4564 4565 4566 4567 4568 4569 4570 4571 4572 4573 4574 4575 4576 4577 4578 4579 4580 4581 4582 4583 4584 4585 4586 4587 4588 4589 4590 4591 4592 4593 4594 4595 4596 4597 4598 4599 4600 4601 4602 4603 4604 4605 4606 4607 4608 4609 4610 4611 4612 4613 4614 4615 4616 4617 4618 4619 4620 4621 4622 4623 4624 4625 4626 4627 4628 4629 4630 4631 4632 4633 4634 4635 4636 4637 4638 4639 4640 4641 4642 4643 4644 4645 4646 4647 4648 4649 4650 4651 4652 4653 4654 4655 4656 4657 4658 4659 4660 4661 4662 4663 4664 4665 4666 4667 4668 4669 4670 4671 4672 4673 4674 4675 4676 4677 4678 4679 4680 4681 4682 4683 4684 4685 4686 4687 4688 4689 4690 4691 4692 4693 4694 4695 4696 4697 4698 4699 4700 4701 4702 4703 4704 4705 4706 4707 4708 4709 4710 4711 4712 4713 4714 4715 4716 4717 4718 4719 4720 4721 4722 4723 4724 4725 4726 4727 4728 4729 4730 4731 4732 4733 4734 4735 4736 4737 4738 4739 4740 4741 4742 4743 4744 4745 4746 4747 4748 4749 4750 4751 4752 4753 4754 4755 4756 4757 4758 4759 4760 4761 4762 4763 4764 4765 4766 4767 4768 4769 4770 4771 4772 4773 4774 4775 4776 4777 4778 4779 4780 4781 4782 4783 4784 4785 4786 4787 4788 4789 4790 4791 4792 4793 4794 4795 4796 4797 4798 4799 4800 4801 4802 4803 4804 4805 4806 4807 4808 4809 4810 4811 4812 4813 4814 4815 4816 4817 4818 4819 4820 4821 4822 4823 4824 4825 4826 4827 4828 4829 4830 4831 4832 4833 4834 4835 4836 4837 4838 4839 4840 4841 4842 4843 4844 4845 4846 4847 4848 4849 4850 4851 4852 4853 4854 4855 4856 4857 4858 4859 4860 4861 4862 4863 4864 4865 4866 4867 4868 4869 4870 4871 4872 4873 4874 4875 4876 4877 4878 4879 4880 4881 4882 4883 4884 4885 4886 4887 4888 4889 4890 4891 4892 4893 4894 4895 4896 4897 4898 4899 4900 4901 4902 4903 4904 4905 4906 4907 4908 4909 4910 4911 4912 4913 4914 4915 4916 4917 4918 4919 4920 4921 4922 4923 4924 4925 4926 4927 4928 4929 4930 4931 4932 4933 4934 4935 4936 4937 4938 4939 4940 4941 4942 4943 4944 4945 4946 4947 4948 4949 4950 4951 4952 4953 4954 4955 4956 4957 4958 4959 4960 4961 4962 4963 4964 4965 4966 4967 4968 4969 4970 4971 4972 4973 4974 4975 4976 4977 4978 4979 4980 4981 4982 4983 4984 4985 4986 4987 4988 4989 4990 4991 4992 4993 4994 4995 4996 4997 4998 4999 5000 5001 5002 5003 5004 5005 5006 5007 5008 5009 5010 5011 5012 5013 5014 5015 5016 5017 5018 5019 5020 5021 5022 5023 5024 5025 5026 5027 5028 5029 5030 5031 5032 5033 5034 5035 5036 5037 5038 5039 5040 5041 5042 5043 5044 5045 5046 5047 5048 5049 5050 5051 5052 5053 5054 5055 5056 5057 5058 5059 5060 5061 5062 5063 5064 5065 5066 5067 5068 5069 5070 5071 5072 5073 5074 5075 5076 5077 5078 5079 5080 5081 5082 5083 5084 5085 5086 5087 5088 5089 5090 5091 5092 5093 5094 5095 5096 5097 5098 5099 5100 5101 5102 5103 5104 5105 5106 5107 5108 5109 5110 5111 5112 5113 5114 5115 5116 5117 5118 5119 5120 5121 5122 5123 5124 5125 5126 5127 5128 5129 5130 5131 5132 5133 5134 5135 5136 5137 5138 5139 5140 5141 5142 5143 5144 5145 5146 5147 5148 5149 5150 5151 5152 5153 5154 5155 5156 5157 5158 5159 5160 5161 5162 5163 5164 5165 5166 5167 5168 5169 5170 5171 5172 5173 5174 5175 5176 5177 5178 5179 5180 5181 5182 5183 5184 5185 5186 5187 5188 5189 5190 5191 5192 5193 5194 5195 5196 5197 5198 5199 5200 5201 5202 5203 5204 5205 5206 5207 5208 5209 5210 5211 5212 5213 5214 5215 5216 5217 5218 5219 5220 5221 5222 5223 5224 5225 5226 5227 5228 5229 5230 5231 5232 5233 5234 5235 5236 5237 5238 5239 5240 5241 5242 5243 5244 5245 5246 5247 5248 5249 5250 5251 5252 5253 5254 5255 5256 5257 5258 5259 5260 5261 5262 5263 5264 5265 5266 5267 5268 5269 5270 5271 5272 5273 5274 5275 5276 5277 5278 5279 5280 5281 5282 5283 5284 5285 5286 5287 5288 5289 5290 5291 5292 5293 5294 5295 5296 5297 5298 5299 5300 5301 5302 5303 5304 5305 5306 5307 5308 5309 5310 5311 5312 5313 5314 5315 5316 5317 5318 5319 5320 5321 5322 5323 5324 5325 5326 5327 5328 5329 5330 5331 5332 5333 5334 5335 5336 5337 5338 5339 5340 5341 5342 5343 5344 5345 5346 5347 5348 5349 5350 5351 5352 5353 5354 5355 5356 5357 5358 5359 5360 5361 5362 5363 5364 5365 5366 5367 5368 5369 5370 5371 5372 5373 5374 5375 5376 5377 5378 5379 5380 5381 5382 5383 5384 5385 5386 5387 5388 5389 5390 5391 5392 5393 5394 5395 5396 5397 5398 5399 5400 5401 5402 5403 5404 5405 5406 5407 5408 5409 5410 5411 5412 5413 5414 5415 5416 5417 5418 5419 5420 5421 5422 5423 5424 5425 5426 5427 5428 5429 5430 5431 5432 5433 5434 5435 5436 5437 5438 5439 5440 5441 5442 5443 5444 5445 5446 5447 5448 5449 5450 5451 5452 5453 5454 5455 5456 5457 5458 5459 5460 5461 5462 5463 5464 5465 5466 5467 5468 5469 5470 5471 5472 5473 5474 5475 5476 5477 5478 5479 5480 5481 5482 5483 5484 5485 5486 5487 5488 5489 5490 5491 5492 5493 5494 5495 5496 5497 5498 5499 5500 5501 5502 5503 5504 5505 5506 5507 5508 5509 5510 5511 5512 5513 5514 5515 5516 5517 5518 5519 5520 5521 5522 5523 5524 5525 5526 5527 5528 5529 5530 5531 5532 5533 5534 5535 5536 5537 5538 5539 5540 5541 5542 5543 5544 5545 5546 5547 5548 5549 5550 5551 5552 5553 5554 5555 5556 5557 5558 5559 5560 5561 5562 5563 5564 5565 5566 5567 5568 5569 5570 5571 5572 5573 5574 5575 5576 5577 5578 5579 5580 5581 5582 5583 5584 5585 5586 5587 5588 5589 5590 5591 5592 5593 5594 5595 5596 5597 5598 5599 5600 5601 5602 5603 5604 5605 5606 5607 5608 5609 5610 5611 5612 5613 5614 5615 5616 5617 5618 5619 5620 5621 5622 5623 5624 5625 5626 5627 5628 5629 5630 5631 5632 5633 5634 5635 5636 5637 5638 5639 5640 5641 5642 5643 5644 5645 5646 5647 5648 5649 5650 5651 5652 5653 5654 5655 5656 5657 5658 5659 5660 5661 5662 5663 5664 5665 5666 5667 5668 5669 5670 5671 5672 5673 5674 5675 5676 5677 5678 5679 5680 5681 5682 5683 5684 5685 5686 5687 5688
// ================================================================= // // * WARNING * // // This file is generated! // // Changes made to this file will be overwritten. If changes are // required to the generated code, the service_crategen project // must be updated to generate the changes. // // ================================================================= use std::error::Error; use std::fmt; #[allow(warnings)] use futures::future; use futures::Future; use rusoto_core::credential::ProvideAwsCredentials; use rusoto_core::region; use rusoto_core::request::{BufferedHttpResponse, DispatchSignedRequest}; use rusoto_core::{Client, RusotoError, RusotoFuture}; use rusoto_core::proto; use rusoto_core::signature::SignedRequest; use serde_json; /// <p>Represents an attribute for describing the key schema for the table and indexes.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct AttributeDefinition { /// <p>A name for the attribute.</p> #[serde(rename = "AttributeName")] pub attribute_name: String, /// <p><p>The data type for the attribute, where:</p> <ul> <li> <p> <code>S</code> - the attribute is of type String</p> </li> <li> <p> <code>N</code> - the attribute is of type Number</p> </li> <li> <p> <code>B</code> - the attribute is of type Binary</p> </li> </ul></p> #[serde(rename = "AttributeType")] pub attribute_type: String, } /// <p>Represents the data for an attribute.</p> <p>Each attribute value is described as a name-value pair. The name is the data type, and the value is the data itself.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/HowItWorks.NamingRulesDataTypes.html#HowItWorks.DataTypes">Data Types</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct AttributeValue { /// <p>An attribute of type Binary. For example:</p> <p> <code>"B": "dGhpcyB0ZXh0IGlzIGJhc2U2NC1lbmNvZGVk"</code> </p> #[serde(rename = "B")] #[serde( deserialize_with = "::rusoto_core::serialization::SerdeBlob::deserialize_blob", serialize_with = "::rusoto_core::serialization::SerdeBlob::serialize_blob", default )] #[serde(skip_serializing_if = "Option::is_none")] pub b: Option<bytes::Bytes>, /// <p>An attribute of type Boolean. For example:</p> <p> <code>"BOOL": true</code> </p> #[serde(rename = "BOOL")] #[serde(skip_serializing_if = "Option::is_none")] pub bool: Option<bool>, /// <p>An attribute of type Binary Set. For example:</p> <p> <code>"BS": ["U3Vubnk=", "UmFpbnk=", "U25vd3k="]</code> </p> #[serde(rename = "BS")] #[serde( deserialize_with = "::rusoto_core::serialization::SerdeBlobList::deserialize_blob_list", serialize_with = "::rusoto_core::serialization::SerdeBlobList::serialize_blob_list", default )] #[serde(skip_serializing_if = "Option::is_none")] pub bs: Option<Vec<bytes::Bytes>>, /// <p>An attribute of type List. For example:</p> <p> <code>"L": [ {"S": "Cookies"} , {"S": "Coffee"}, {"N", "3.14159"}]</code> </p> #[serde(rename = "L")] #[serde(skip_serializing_if = "Option::is_none")] pub l: Option<Vec<AttributeValue>>, /// <p>An attribute of type Map. For example:</p> <p> <code>"M": {"Name": {"S": "Joe"}, "Age": {"N": "35"}}</code> </p> #[serde(rename = "M")] #[serde(skip_serializing_if = "Option::is_none")] pub m: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>An attribute of type Number. For example:</p> <p> <code>"N": "123.45"</code> </p> <p>Numbers are sent across the network to DynamoDB as strings, to maximize compatibility across languages and libraries. However, DynamoDB treats them as number type attributes for mathematical operations.</p> #[serde(rename = "N")] #[serde(skip_serializing_if = "Option::is_none")] pub n: Option<String>, /// <p>An attribute of type Number Set. For example:</p> <p> <code>"NS": ["42.2", "-19", "7.5", "3.14"]</code> </p> <p>Numbers are sent across the network to DynamoDB as strings, to maximize compatibility across languages and libraries. However, DynamoDB treats them as number type attributes for mathematical operations.</p> #[serde(rename = "NS")] #[serde(skip_serializing_if = "Option::is_none")] pub ns: Option<Vec<String>>, /// <p>An attribute of type Null. For example:</p> <p> <code>"NULL": true</code> </p> #[serde(rename = "NULL")] #[serde(skip_serializing_if = "Option::is_none")] pub null: Option<bool>, /// <p>An attribute of type String. For example:</p> <p> <code>"S": "Hello"</code> </p> #[serde(rename = "S")] #[serde(skip_serializing_if = "Option::is_none")] pub s: Option<String>, /// <p>An attribute of type String Set. For example:</p> <p> <code>"SS": ["Giraffe", "Hippo" ,"Zebra"]</code> </p> #[serde(rename = "SS")] #[serde(skip_serializing_if = "Option::is_none")] pub ss: Option<Vec<String>>, } /// <p>For the <code>UpdateItem</code> operation, represents the attributes to be modified, the action to perform on each, and the new value for each.</p> <note> <p>You cannot use <code>UpdateItem</code> to update any primary key attributes. Instead, you will need to delete the item, and then use <code>PutItem</code> to create a new item with new attributes.</p> </note> <p>Attribute values cannot be null; string and binary type attributes must have lengths greater than zero; and set type attributes must not be empty. Requests with empty values will be rejected with a <code>ValidationException</code> exception.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct AttributeValueUpdate { /// <p><p>Specifies how to perform the update. Valid values are <code>PUT</code> (default), <code>DELETE</code>, and <code>ADD</code>. The behavior depends on whether the specified primary key already exists in the table.</p> <p> <b>If an item with the specified <i>Key</i> is found in the table:</b> </p> <ul> <li> <p> <code>PUT</code> - Adds the specified attribute to the item. If the attribute already exists, it is replaced by the new value. </p> </li> <li> <p> <code>DELETE</code> - If no value is specified, the attribute and its value are removed from the item. The data type of the specified value must match the existing value's data type.</p> <p>If a <i>set</i> of values is specified, then those values are subtracted from the old set. For example, if the attribute value was the set <code>[a,b,c]</code> and the <code>DELETE</code> action specified <code>[a,c]</code>, then the final attribute value would be <code>[b]</code>. Specifying an empty set is an error.</p> </li> <li> <p> <code>ADD</code> - If the attribute does not already exist, then the attribute and its values are added to the item. If the attribute does exist, then the behavior of <code>ADD</code> depends on the data type of the attribute:</p> <ul> <li> <p>If the existing attribute is a number, and if <code>Value</code> is also a number, then the <code>Value</code> is mathematically added to the existing attribute. If <code>Value</code> is a negative number, then it is subtracted from the existing attribute.</p> <note> <p> If you use <code>ADD</code> to increment or decrement a number value for an item that doesn't exist before the update, DynamoDB uses 0 as the initial value.</p> <p>In addition, if you use <code>ADD</code> to update an existing item, and intend to increment or decrement an attribute value which does not yet exist, DynamoDB uses <code>0</code> as the initial value. For example, suppose that the item you want to update does not yet have an attribute named <i>itemcount</i>, but you decide to <code>ADD</code> the number <code>3</code> to this attribute anyway, even though it currently does not exist. DynamoDB will create the <i>itemcount</i> attribute, set its initial value to <code>0</code>, and finally add <code>3</code> to it. The result will be a new <i>itemcount</i> attribute in the item, with a value of <code>3</code>.</p> </note> </li> <li> <p>If the existing data type is a set, and if the <code>Value</code> is also a set, then the <code>Value</code> is added to the existing set. (This is a <i>set</i> operation, not mathematical addition.) For example, if the attribute value was the set <code>[1,2]</code>, and the <code>ADD</code> action specified <code>[3]</code>, then the final attribute value would be <code>[1,2,3]</code>. An error occurs if an Add action is specified for a set attribute and the attribute type specified does not match the existing set type. </p> <p>Both sets must have the same primitive data type. For example, if the existing data type is a set of strings, the <code>Value</code> must also be a set of strings. The same holds true for number sets and binary sets.</p> </li> </ul> <p>This action is only valid for an existing attribute whose data type is number or is a set. Do not use <code>ADD</code> for any other data types.</p> </li> </ul> <p> <b>If no item with the specified <i>Key</i> is found:</b> </p> <ul> <li> <p> <code>PUT</code> - DynamoDB creates a new item with the specified primary key, and then adds the attribute. </p> </li> <li> <p> <code>DELETE</code> - Nothing happens; there is no attribute to delete.</p> </li> <li> <p> <code>ADD</code> - DynamoDB creates an item with the supplied primary key and number (or set of numbers) for the attribute value. The only data types allowed are number and number set; no other data types can be specified.</p> </li> </ul></p> #[serde(rename = "Action")] #[serde(skip_serializing_if = "Option::is_none")] pub action: Option<String>, /// <p>Represents the data for an attribute.</p> <p>Each attribute value is described as a name-value pair. The name is the data type, and the value is the data itself.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/HowItWorks.NamingRulesDataTypes.html#HowItWorks.DataTypes">Data Types</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> #[serde(rename = "Value")] #[serde(skip_serializing_if = "Option::is_none")] pub value: Option<AttributeValue>, } /// <p>Represents the properties of the scaling policy.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct AutoScalingPolicyDescription { /// <p>The name of the scaling policy.</p> #[serde(rename = "PolicyName")] #[serde(skip_serializing_if = "Option::is_none")] pub policy_name: Option<String>, /// <p>Represents a target tracking scaling policy configuration.</p> #[serde(rename = "TargetTrackingScalingPolicyConfiguration")] #[serde(skip_serializing_if = "Option::is_none")] pub target_tracking_scaling_policy_configuration: Option<AutoScalingTargetTrackingScalingPolicyConfigurationDescription>, } /// <p>Represents the autoscaling policy to be modified.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct AutoScalingPolicyUpdate { /// <p>The name of the scaling policy.</p> #[serde(rename = "PolicyName")] #[serde(skip_serializing_if = "Option::is_none")] pub policy_name: Option<String>, /// <p>Represents a target tracking scaling policy configuration.</p> #[serde(rename = "TargetTrackingScalingPolicyConfiguration")] pub target_tracking_scaling_policy_configuration: AutoScalingTargetTrackingScalingPolicyConfigurationUpdate, } /// <p>Represents the autoscaling settings for a global table or global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct AutoScalingSettingsDescription { /// <p>Disabled autoscaling for this global table or global secondary index.</p> #[serde(rename = "AutoScalingDisabled")] #[serde(skip_serializing_if = "Option::is_none")] pub auto_scaling_disabled: Option<bool>, /// <p>Role ARN used for configuring autoScaling policy.</p> #[serde(rename = "AutoScalingRoleArn")] #[serde(skip_serializing_if = "Option::is_none")] pub auto_scaling_role_arn: Option<String>, /// <p>The maximum capacity units that a global table or global secondary index should be scaled up to.</p> #[serde(rename = "MaximumUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub maximum_units: Option<i64>, /// <p>The minimum capacity units that a global table or global secondary index should be scaled down to.</p> #[serde(rename = "MinimumUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub minimum_units: Option<i64>, /// <p>Information about the scaling policies.</p> #[serde(rename = "ScalingPolicies")] #[serde(skip_serializing_if = "Option::is_none")] pub scaling_policies: Option<Vec<AutoScalingPolicyDescription>>, } /// <p>Represents the autoscaling settings to be modified for a global table or global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct AutoScalingSettingsUpdate { /// <p>Disabled autoscaling for this global table or global secondary index.</p> #[serde(rename = "AutoScalingDisabled")] #[serde(skip_serializing_if = "Option::is_none")] pub auto_scaling_disabled: Option<bool>, /// <p>Role ARN used for configuring autoscaling policy.</p> #[serde(rename = "AutoScalingRoleArn")] #[serde(skip_serializing_if = "Option::is_none")] pub auto_scaling_role_arn: Option<String>, /// <p>The maximum capacity units that a global table or global secondary index should be scaled up to.</p> #[serde(rename = "MaximumUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub maximum_units: Option<i64>, /// <p>The minimum capacity units that a global table or global secondary index should be scaled down to.</p> #[serde(rename = "MinimumUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub minimum_units: Option<i64>, /// <p>The scaling policy to apply for scaling target global table or global secondary index capacity units.</p> #[serde(rename = "ScalingPolicyUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub scaling_policy_update: Option<AutoScalingPolicyUpdate>, } /// <p>Represents the properties of a target tracking scaling policy.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct AutoScalingTargetTrackingScalingPolicyConfigurationDescription { /// <p>Indicates whether scale in by the target tracking policy is disabled. If the value is true, scale in is disabled and the target tracking policy won't remove capacity from the scalable resource. Otherwise, scale in is enabled and the target tracking policy can remove capacity from the scalable resource. The default value is false.</p> #[serde(rename = "DisableScaleIn")] #[serde(skip_serializing_if = "Option::is_none")] pub disable_scale_in: Option<bool>, /// <p>The amount of time, in seconds, after a scale in activity completes before another scale in activity can start. The cooldown period is used to block subsequent scale in requests until it has expired. You should scale in conservatively to protect your application's availability. However, if another alarm triggers a scale out policy during the cooldown period after a scale-in, application autoscaling scales out your scalable target immediately. </p> #[serde(rename = "ScaleInCooldown")] #[serde(skip_serializing_if = "Option::is_none")] pub scale_in_cooldown: Option<i64>, /// <p>The amount of time, in seconds, after a scale out activity completes before another scale out activity can start. While the cooldown period is in effect, the capacity that has been added by the previous scale out event that initiated the cooldown is calculated as part of the desired capacity for the next scale out. You should continuously (but not excessively) scale out.</p> #[serde(rename = "ScaleOutCooldown")] #[serde(skip_serializing_if = "Option::is_none")] pub scale_out_cooldown: Option<i64>, /// <p>The target value for the metric. The range is 8.515920e-109 to 1.174271e+108 (Base 10) or 2e-360 to 2e360 (Base 2).</p> #[serde(rename = "TargetValue")] pub target_value: f64, } /// <p>Represents the settings of a target tracking scaling policy that will be modified.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct AutoScalingTargetTrackingScalingPolicyConfigurationUpdate { /// <p>Indicates whether scale in by the target tracking policy is disabled. If the value is true, scale in is disabled and the target tracking policy won't remove capacity from the scalable resource. Otherwise, scale in is enabled and the target tracking policy can remove capacity from the scalable resource. The default value is false.</p> #[serde(rename = "DisableScaleIn")] #[serde(skip_serializing_if = "Option::is_none")] pub disable_scale_in: Option<bool>, /// <p>The amount of time, in seconds, after a scale in activity completes before another scale in activity can start. The cooldown period is used to block subsequent scale in requests until it has expired. You should scale in conservatively to protect your application's availability. However, if another alarm triggers a scale out policy during the cooldown period after a scale-in, application autoscaling scales out your scalable target immediately. </p> #[serde(rename = "ScaleInCooldown")] #[serde(skip_serializing_if = "Option::is_none")] pub scale_in_cooldown: Option<i64>, /// <p>The amount of time, in seconds, after a scale out activity completes before another scale out activity can start. While the cooldown period is in effect, the capacity that has been added by the previous scale out event that initiated the cooldown is calculated as part of the desired capacity for the next scale out. You should continuously (but not excessively) scale out.</p> #[serde(rename = "ScaleOutCooldown")] #[serde(skip_serializing_if = "Option::is_none")] pub scale_out_cooldown: Option<i64>, /// <p>The target value for the metric. The range is 8.515920e-109 to 1.174271e+108 (Base 10) or 2e-360 to 2e360 (Base 2).</p> #[serde(rename = "TargetValue")] pub target_value: f64, } /// <p>Contains the description of the backup created for the table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BackupDescription { /// <p>Contains the details of the backup created for the table. </p> #[serde(rename = "BackupDetails")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_details: Option<BackupDetails>, /// <p>Contains the details of the table when the backup was created. </p> #[serde(rename = "SourceTableDetails")] #[serde(skip_serializing_if = "Option::is_none")] pub source_table_details: Option<SourceTableDetails>, /// <p>Contains the details of the features enabled on the table when the backup was created. For example, LSIs, GSIs, streams, TTL.</p> #[serde(rename = "SourceTableFeatureDetails")] #[serde(skip_serializing_if = "Option::is_none")] pub source_table_feature_details: Option<SourceTableFeatureDetails>, } /// <p>Contains the details of the backup created for the table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BackupDetails { /// <p>ARN associated with the backup.</p> #[serde(rename = "BackupArn")] pub backup_arn: String, /// <p>Time at which the backup was created. This is the request time of the backup. </p> #[serde(rename = "BackupCreationDateTime")] pub backup_creation_date_time: f64, /// <p>Time at which the automatic on-demand backup created by DynamoDB will expire. This <code>SYSTEM</code> on-demand backup expires automatically 35 days after its creation.</p> #[serde(rename = "BackupExpiryDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_expiry_date_time: Option<f64>, /// <p>Name of the requested backup.</p> #[serde(rename = "BackupName")] pub backup_name: String, /// <p>Size of the backup in bytes.</p> #[serde(rename = "BackupSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_size_bytes: Option<i64>, /// <p>Backup can be in one of the following states: CREATING, ACTIVE, DELETED. </p> #[serde(rename = "BackupStatus")] pub backup_status: String, /// <p><p>BackupType:</p> <ul> <li> <p> <code>USER</code> - You create and manage these using the on-demand backup feature.</p> </li> <li> <p> <code>SYSTEM</code> - If you delete a table with point-in-time recovery enabled, a <code>SYSTEM</code> backup is automatically created and is retained for 35 days (at no additional cost). System backups allow you to restore the deleted table to the state it was in just before the point of deletion. </p> </li> <li> <p> <code>AWS_BACKUP</code> - On-demand backup created by you from AWS Backup service.</p> </li> </ul></p> #[serde(rename = "BackupType")] pub backup_type: String, } /// <p>Contains details for the backup.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BackupSummary { /// <p>ARN associated with the backup.</p> #[serde(rename = "BackupArn")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_arn: Option<String>, /// <p>Time at which the backup was created.</p> #[serde(rename = "BackupCreationDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_creation_date_time: Option<f64>, /// <p>Time at which the automatic on-demand backup created by DynamoDB will expire. This <code>SYSTEM</code> on-demand backup expires automatically 35 days after its creation.</p> #[serde(rename = "BackupExpiryDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_expiry_date_time: Option<f64>, /// <p>Name of the specified backup.</p> #[serde(rename = "BackupName")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_name: Option<String>, /// <p>Size of the backup in bytes.</p> #[serde(rename = "BackupSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_size_bytes: Option<i64>, /// <p>Backup can be in one of the following states: CREATING, ACTIVE, DELETED.</p> #[serde(rename = "BackupStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_status: Option<String>, /// <p><p>BackupType:</p> <ul> <li> <p> <code>USER</code> - You create and manage these using the on-demand backup feature.</p> </li> <li> <p> <code>SYSTEM</code> - If you delete a table with point-in-time recovery enabled, a <code>SYSTEM</code> backup is automatically created and is retained for 35 days (at no additional cost). System backups allow you to restore the deleted table to the state it was in just before the point of deletion. </p> </li> <li> <p> <code>AWS_BACKUP</code> - On-demand backup created by you from AWS Backup service.</p> </li> </ul></p> #[serde(rename = "BackupType")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_type: Option<String>, /// <p>ARN associated with the table.</p> #[serde(rename = "TableArn")] #[serde(skip_serializing_if = "Option::is_none")] pub table_arn: Option<String>, /// <p>Unique identifier for the table.</p> #[serde(rename = "TableId")] #[serde(skip_serializing_if = "Option::is_none")] pub table_id: Option<String>, /// <p>Name of the table.</p> #[serde(rename = "TableName")] #[serde(skip_serializing_if = "Option::is_none")] pub table_name: Option<String>, } /// <p>Represents the input of a <code>BatchGetItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct BatchGetItemInput { /// <p><p>A map of one or more table names and, for each table, a map that describes one or more items to retrieve from that table. Each table name can be used only once per <code>BatchGetItem</code> request.</p> <p>Each element in the map of items to retrieve consists of the following:</p> <ul> <li> <p> <code>ConsistentRead</code> - If <code>true</code>, a strongly consistent read is used; if <code>false</code> (the default), an eventually consistent read is used.</p> </li> <li> <p> <code>ExpressionAttributeNames</code> - One or more substitution tokens for attribute names in the <code>ProjectionExpression</code> parameter. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </li> <li> <p> <code>Keys</code> - An array of primary key attribute values that define specific items in the table. For each primary key, you must provide <i>all</i> of the key attributes. For example, with a simple primary key, you only need to provide the partition key value. For a composite key, you must provide <i>both</i> the partition key value and the sort key value.</p> </li> <li> <p> <code>ProjectionExpression</code> - A string that identifies one or more attributes to retrieve from the table. These attributes can include scalars, sets, or elements of a JSON document. The attributes in the expression must be separated by commas.</p> <p>If no attribute names are specified, then all attributes will be returned. If any of the requested attributes are not found, they will not appear in the result.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </li> <li> <p> <code>AttributesToGet</code> - This is a legacy parameter. Use <code>ProjectionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.AttributesToGet.html">AttributesToGet</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> </li> </ul></p> #[serde(rename = "RequestItems")] pub request_items: ::std::collections::HashMap<String, KeysAndAttributes>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, } /// <p>Represents the output of a <code>BatchGetItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BatchGetItemOutput { /// <p><p>The read capacity units consumed by the entire <code>BatchGetItem</code> operation.</p> <p>Each element consists of:</p> <ul> <li> <p> <code>TableName</code> - The table that consumed the provisioned throughput.</p> </li> <li> <p> <code>CapacityUnits</code> - The total number of capacity units consumed.</p> </li> </ul></p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<Vec<ConsumedCapacity>>, /// <p>A map of table name to a list of items. Each object in <code>Responses</code> consists of a table name, along with a map of attribute data consisting of the data type and attribute value.</p> #[serde(rename = "Responses")] #[serde(skip_serializing_if = "Option::is_none")] pub responses: Option< ::std::collections::HashMap< String, Vec<::std::collections::HashMap<String, AttributeValue>>, >, >, /// <p>A map of tables and their respective keys that were not processed with the current response. The <code>UnprocessedKeys</code> value is in the same form as <code>RequestItems</code>, so the value can be provided directly to a subsequent <code>BatchGetItem</code> operation. For more information, see <code>RequestItems</code> in the Request Parameters section.</p> <p>Each element consists of:</p> <ul> <li> <p> <code>Keys</code> - An array of primary key attribute values that define specific items in the table.</p> </li> <li> <p> <code>ProjectionExpression</code> - One or more attributes to be retrieved from the table or index. By default, all attributes are returned. If a requested attribute is not found, it does not appear in the result.</p> </li> <li> <p> <code>ConsistentRead</code> - The consistency of a read operation. If set to <code>true</code>, then a strongly consistent read is used; otherwise, an eventually consistent read is used.</p> </li> </ul> <p>If there are no unprocessed keys remaining, the response contains an empty <code>UnprocessedKeys</code> map.</p> #[serde(rename = "UnprocessedKeys")] #[serde(skip_serializing_if = "Option::is_none")] pub unprocessed_keys: Option<::std::collections::HashMap<String, KeysAndAttributes>>, } /// <p>Represents the input of a <code>BatchWriteItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct BatchWriteItemInput { /// <p><p>A map of one or more table names and, for each table, a list of operations to be performed (<code>DeleteRequest</code> or <code>PutRequest</code>). Each element in the map consists of the following:</p> <ul> <li> <p> <code>DeleteRequest</code> - Perform a <code>DeleteItem</code> operation on the specified item. The item to be deleted is identified by a <code>Key</code> subelement:</p> <ul> <li> <p> <code>Key</code> - A map of primary key attribute values that uniquely identify the item. Each entry in this map consists of an attribute name and an attribute value. For each primary key, you must provide <i>all</i> of the key attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for <i>both</i> the partition key and the sort key.</p> </li> </ul> </li> <li> <p> <code>PutRequest</code> - Perform a <code>PutItem</code> operation on the specified item. The item to be put is identified by an <code>Item</code> subelement:</p> <ul> <li> <p> <code>Item</code> - A map of attributes and their values. Each entry in this map consists of an attribute name and an attribute value. Attribute values must not be null; string and binary type attributes must have lengths greater than zero; and set type attributes must not be empty. Requests that contain empty values will be rejected with a <code>ValidationException</code> exception.</p> <p>If you specify any attributes that are part of an index key, then the data types for those attributes must match those of the schema in the table's attribute definition.</p> </li> </ul> </li> </ul></p> #[serde(rename = "RequestItems")] pub request_items: ::std::collections::HashMap<String, Vec<WriteRequest>>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Determines whether item collection metrics are returned. If set to <code>SIZE</code>, the response includes statistics about item collections, if any, that were modified during the operation are returned in the response. If set to <code>NONE</code> (the default), no statistics are returned.</p> #[serde(rename = "ReturnItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub return_item_collection_metrics: Option<String>, } /// <p>Represents the output of a <code>BatchWriteItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BatchWriteItemOutput { /// <p><p>The capacity units consumed by the entire <code>BatchWriteItem</code> operation.</p> <p>Each element consists of:</p> <ul> <li> <p> <code>TableName</code> - The table that consumed the provisioned throughput.</p> </li> <li> <p> <code>CapacityUnits</code> - The total number of capacity units consumed.</p> </li> </ul></p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<Vec<ConsumedCapacity>>, /// <p><p>A list of tables that were processed by <code>BatchWriteItem</code> and, for each table, information about any item collections that were affected by individual <code>DeleteItem</code> or <code>PutItem</code> operations.</p> <p>Each entry consists of the following subelements:</p> <ul> <li> <p> <code>ItemCollectionKey</code> - The partition key value of the item collection. This is the same as the partition key value of the item.</p> </li> <li> <p> <code>SizeEstimateRangeGB</code> - An estimate of item collection size, expressed in GB. This is a two-element array containing a lower bound and an upper bound for the estimate. The estimate includes the size of all the items in the table, plus the size of all attributes projected into all of the local secondary indexes on the table. Use this estimate to measure whether a local secondary index is approaching its size limit.</p> <p>The estimate is subject to change over time; therefore, do not rely on the precision or accuracy of the estimate.</p> </li> </ul></p> #[serde(rename = "ItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_metrics: Option<::std::collections::HashMap<String, Vec<ItemCollectionMetrics>>>, /// <p>A map of tables and requests against those tables that were not processed. The <code>UnprocessedItems</code> value is in the same form as <code>RequestItems</code>, so you can provide this value directly to a subsequent <code>BatchGetItem</code> operation. For more information, see <code>RequestItems</code> in the Request Parameters section.</p> <p>Each <code>UnprocessedItems</code> entry consists of a table name and, for that table, a list of operations to perform (<code>DeleteRequest</code> or <code>PutRequest</code>).</p> <ul> <li> <p> <code>DeleteRequest</code> - Perform a <code>DeleteItem</code> operation on the specified item. The item to be deleted is identified by a <code>Key</code> subelement:</p> <ul> <li> <p> <code>Key</code> - A map of primary key attribute values that uniquely identify the item. Each entry in this map consists of an attribute name and an attribute value.</p> </li> </ul> </li> <li> <p> <code>PutRequest</code> - Perform a <code>PutItem</code> operation on the specified item. The item to be put is identified by an <code>Item</code> subelement:</p> <ul> <li> <p> <code>Item</code> - A map of attributes and their values. Each entry in this map consists of an attribute name and an attribute value. Attribute values must not be null; string and binary type attributes must have lengths greater than zero; and set type attributes must not be empty. Requests that contain empty values will be rejected with a <code>ValidationException</code> exception.</p> <p>If you specify any attributes that are part of an index key, then the data types for those attributes must match those of the schema in the table's attribute definition.</p> </li> </ul> </li> </ul> <p>If there are no unprocessed items remaining, the response contains an empty <code>UnprocessedItems</code> map.</p> #[serde(rename = "UnprocessedItems")] #[serde(skip_serializing_if = "Option::is_none")] pub unprocessed_items: Option<::std::collections::HashMap<String, Vec<WriteRequest>>>, } /// <p>Contains the details for the read/write capacity mode.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct BillingModeSummary { /// <p><p>Controls how you are charged for read and write throughput and how you manage capacity. This setting can be changed later.</p> <ul> <li> <p> <code>PROVISIONED</code> - Sets the read/write capacity mode to <code>PROVISIONED</code>. We recommend using <code>PROVISIONED</code> for predictable workloads.</p> </li> <li> <p> <code>PAY<em>PER</em>REQUEST</code> - Sets the read/write capacity mode to <code>PAY<em>PER</em>REQUEST</code>. We recommend using <code>PAY<em>PER</em>REQUEST</code> for unpredictable workloads. </p> </li> </ul></p> #[serde(rename = "BillingMode")] #[serde(skip_serializing_if = "Option::is_none")] pub billing_mode: Option<String>, /// <p>Represents the time when <code>PAY_PER_REQUEST</code> was last set as the read/write capacity mode.</p> #[serde(rename = "LastUpdateToPayPerRequestDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub last_update_to_pay_per_request_date_time: Option<f64>, } /// <p>An ordered list of errors for each item in the request which caused the transaction to get cancelled. The values of the list are ordered according to the ordering of the <code>TransactWriteItems</code> request parameter. If no error occurred for the associated item an error with a Null code and Null message will be present. </p> #[derive(Default, Debug, Clone, PartialEq)] pub struct CancellationReason { /// <p>Status code for the result of the cancelled transaction.</p> pub code: Option<String>, /// <p>Item in the request which caused the transaction to get cancelled.</p> pub item: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>Cancellation reason message description.</p> pub message: Option<String>, } /// <p>Represents the amount of provisioned throughput capacity consumed on a table or an index.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct Capacity { /// <p>The total number of capacity units consumed on a table or an index.</p> #[serde(rename = "CapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub capacity_units: Option<f64>, /// <p>The total number of read capacity units consumed on a table or an index.</p> #[serde(rename = "ReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub read_capacity_units: Option<f64>, /// <p>The total number of write capacity units consumed on a table or an index.</p> #[serde(rename = "WriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub write_capacity_units: Option<f64>, } /// <p><p>Represents the selection criteria for a <code>Query</code> or <code>Scan</code> operation:</p> <ul> <li> <p>For a <code>Query</code> operation, <code>Condition</code> is used for specifying the <code>KeyConditions</code> to use when querying a table or an index. For <code>KeyConditions</code>, only the following comparison operators are supported:</p> <p> <code>EQ | LE | LT | GE | GT | BEGINS_WITH | BETWEEN</code> </p> <p> <code>Condition</code> is also used in a <code>QueryFilter</code>, which evaluates the query results and returns only the desired values.</p> </li> <li> <p>For a <code>Scan</code> operation, <code>Condition</code> is used in a <code>ScanFilter</code>, which evaluates the scan results and returns only the desired values.</p> </li> </ul></p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct Condition { /// <p>One or more values to evaluate against the supplied attribute. The number of values in the list depends on the <code>ComparisonOperator</code> being used.</p> <p>For type Number, value comparisons are numeric.</p> <p>String value comparisons for greater than, equals, or less than are based on ASCII character code values. For example, <code>a</code> is greater than <code>A</code>, and <code>a</code> is greater than <code>B</code>. For a list of code values, see <a href="http://en.wikipedia.org/wiki/ASCII#ASCII_printable_characters">http://en.wikipedia.org/wiki/ASCII#ASCII_printable_characters</a>.</p> <p>For Binary, DynamoDB treats each byte of the binary data as unsigned when it compares binary values.</p> #[serde(rename = "AttributeValueList")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_value_list: Option<Vec<AttributeValue>>, /// <p>A comparator for evaluating attributes. For example, equals, greater than, less than, etc.</p> <p>The following comparison operators are available:</p> <p> <code>EQ | NE | LE | LT | GE | GT | NOT_NULL | NULL | CONTAINS | NOT_CONTAINS | BEGINS_WITH | IN | BETWEEN</code> </p> <p>The following are descriptions of each comparison operator.</p> <ul> <li> <p> <code>EQ</code> : Equal. <code>EQ</code> is supported for all data types, including lists and maps.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, Binary, String Set, Number Set, or Binary Set. If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not equal <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>NE</code> : Not equal. <code>NE</code> is supported for all data types, including lists and maps.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String, Number, Binary, String Set, Number Set, or Binary Set. If an item contains an <code>AttributeValue</code> of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not equal <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>LE</code> : Less than or equal. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>LT</code> : Less than. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>GE</code> : Greater than or equal. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>GT</code> : Greater than. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>NOT_NULL</code> : The attribute exists. <code>NOT_NULL</code> is supported for all data types, including lists and maps.</p> <note> <p>This operator tests for the existence of an attribute, not its data type. If the data type of attribute "<code>a</code>" is null, and you evaluate it using <code>NOT_NULL</code>, the result is a Boolean <code>true</code>. This result is because the attribute "<code>a</code>" exists; its data type is not relevant to the <code>NOT_NULL</code> comparison operator.</p> </note> </li> <li> <p> <code>NULL</code> : The attribute does not exist. <code>NULL</code> is supported for all data types, including lists and maps.</p> <note> <p>This operator tests for the nonexistence of an attribute, not its data type. If the data type of attribute "<code>a</code>" is null, and you evaluate it using <code>NULL</code>, the result is a Boolean <code>false</code>. This is because the attribute "<code>a</code>" exists; its data type is not relevant to the <code>NULL</code> comparison operator.</p> </note> </li> <li> <p> <code>CONTAINS</code> : Checks for a subsequence, or value in a set.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If the target attribute of the comparison is of type String, then the operator checks for a substring match. If the target attribute of the comparison is of type Binary, then the operator looks for a subsequence of the target that matches the input. If the target attribute of the comparison is a set ("<code>SS</code>", "<code>NS</code>", or "<code>BS</code>"), then the operator evaluates to true if it finds an exact match with any member of the set.</p> <p>CONTAINS is supported for lists: When evaluating "<code>a CONTAINS b</code>", "<code>a</code>" can be a list; however, "<code>b</code>" cannot be a set, a map, or a list.</p> </li> <li> <p> <code>NOT_CONTAINS</code> : Checks for absence of a subsequence, or absence of a value in a set.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If the target attribute of the comparison is a String, then the operator checks for the absence of a substring match. If the target attribute of the comparison is Binary, then the operator checks for the absence of a subsequence of the target that matches the input. If the target attribute of the comparison is a set ("<code>SS</code>", "<code>NS</code>", or "<code>BS</code>"), then the operator evaluates to true if it <i>does not</i> find an exact match with any member of the set.</p> <p>NOT_CONTAINS is supported for lists: When evaluating "<code>a NOT CONTAINS b</code>", "<code>a</code>" can be a list; however, "<code>b</code>" cannot be a set, a map, or a list.</p> </li> <li> <p> <code>BEGINS_WITH</code> : Checks for a prefix. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String or Binary (not a Number or a set type). The target attribute of the comparison must be of type String or Binary (not a Number or a set type).</p> <p/> </li> <li> <p> <code>IN</code> : Checks for matching elements in a list.</p> <p> <code>AttributeValueList</code> can contain one or more <code>AttributeValue</code> elements of type String, Number, or Binary. These attributes are compared against an existing attribute of an item. If any elements of the input are equal to the item attribute, the expression evaluates to true.</p> </li> <li> <p> <code>BETWEEN</code> : Greater than or equal to the first value, and less than or equal to the second value. </p> <p> <code>AttributeValueList</code> must contain two <code>AttributeValue</code> elements of the same type, either String, Number, or Binary (not a set type). A target attribute matches if the target value is greater than, or equal to, the first element and less than, or equal to, the second element. If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not compare to <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code> </p> </li> </ul> <p>For usage examples of <code>AttributeValueList</code> and <code>ComparisonOperator</code>, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.html">Legacy Conditional Parameters</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ComparisonOperator")] pub comparison_operator: String, } /// <p>Represents a request to perform a check that an item exists or to check the condition of specific attributes of the item..</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ConditionCheck { /// <p>A condition that must be satisfied in order for a conditional update to succeed.</p> #[serde(rename = "ConditionExpression")] pub condition_expression: String, /// <p>One or more substitution tokens for attribute names in an expression.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The primary key of the item to be checked. Each element consists of an attribute name and a value for that attribute.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, /// <p>Use <code>ReturnValuesOnConditionCheckFailure</code> to get the item attributes if the <code>ConditionCheck</code> condition fails. For <code>ReturnValuesOnConditionCheckFailure</code>, the valid values are: NONE and ALL_OLD.</p> #[serde(rename = "ReturnValuesOnConditionCheckFailure")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values_on_condition_check_failure: Option<String>, /// <p>Name of the table for the check item request.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>The capacity units consumed by an operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the request asked for it. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ConsumedCapacity { /// <p>The total number of capacity units consumed by the operation.</p> #[serde(rename = "CapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub capacity_units: Option<f64>, /// <p>The amount of throughput consumed on each global index affected by the operation.</p> #[serde(rename = "GlobalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub global_secondary_indexes: Option<::std::collections::HashMap<String, Capacity>>, /// <p>The amount of throughput consumed on each local index affected by the operation.</p> #[serde(rename = "LocalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub local_secondary_indexes: Option<::std::collections::HashMap<String, Capacity>>, /// <p>The total number of read capacity units consumed by the operation.</p> #[serde(rename = "ReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub read_capacity_units: Option<f64>, /// <p>The amount of throughput consumed on the table affected by the operation.</p> #[serde(rename = "Table")] #[serde(skip_serializing_if = "Option::is_none")] pub table: Option<Capacity>, /// <p>The name of the table that was affected by the operation.</p> #[serde(rename = "TableName")] #[serde(skip_serializing_if = "Option::is_none")] pub table_name: Option<String>, /// <p>The total number of write capacity units consumed by the operation.</p> #[serde(rename = "WriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub write_capacity_units: Option<f64>, } /// <p>Represents the continuous backups and point in time recovery settings on the table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ContinuousBackupsDescription { /// <p> <code>ContinuousBackupsStatus</code> can be one of the following states: ENABLED, DISABLED</p> #[serde(rename = "ContinuousBackupsStatus")] pub continuous_backups_status: String, /// <p>The description of the point in time recovery settings applied to the table.</p> #[serde(rename = "PointInTimeRecoveryDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub point_in_time_recovery_description: Option<PointInTimeRecoveryDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct CreateBackupInput { /// <p>Specified name for the backup.</p> #[serde(rename = "BackupName")] pub backup_name: String, /// <p>The name of the table.</p> #[serde(rename = "TableName")] pub table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct CreateBackupOutput { /// <p>Contains the details of the backup created for the table.</p> #[serde(rename = "BackupDetails")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_details: Option<BackupDetails>, } /// <p>Represents a new global secondary index to be added to an existing table.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct CreateGlobalSecondaryIndexAction { /// <p>The name of the global secondary index to be created.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p>The key schema for the global secondary index.</p> #[serde(rename = "KeySchema")] pub key_schema: Vec<KeySchemaElement>, /// <p>Represents attributes that are copied (projected) from the table into an index. These are in addition to the primary key attributes and index key attributes, which are automatically projected.</p> #[serde(rename = "Projection")] pub projection: Projection, /// <p>Represents the provisioned throughput settings for the specified global secondary index.</p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughput>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct CreateGlobalTableInput { /// <p>The global table name.</p> #[serde(rename = "GlobalTableName")] pub global_table_name: String, /// <p>The regions where the global table needs to be created.</p> #[serde(rename = "ReplicationGroup")] pub replication_group: Vec<Replica>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct CreateGlobalTableOutput { /// <p>Contains the details of the global table.</p> #[serde(rename = "GlobalTableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_description: Option<GlobalTableDescription>, } /// <p>Represents a replica to be added.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct CreateReplicaAction { /// <p>The region of the replica to be added.</p> #[serde(rename = "RegionName")] pub region_name: String, } /// <p>Represents the input of a <code>CreateTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct CreateTableInput { /// <p>An array of attributes that describe the key schema for the table and indexes.</p> #[serde(rename = "AttributeDefinitions")] pub attribute_definitions: Vec<AttributeDefinition>, /// <p><p>Controls how you are charged for read and write throughput and how you manage capacity. This setting can be changed later.</p> <ul> <li> <p> <code>PROVISIONED</code> - Sets the billing mode to <code>PROVISIONED</code>. We recommend using <code>PROVISIONED</code> for predictable workloads.</p> </li> <li> <p> <code>PAY<em>PER</em>REQUEST</code> - Sets the billing mode to <code>PAY<em>PER</em>REQUEST</code>. We recommend using <code>PAY<em>PER</em>REQUEST</code> for unpredictable workloads. </p> </li> </ul></p> #[serde(rename = "BillingMode")] #[serde(skip_serializing_if = "Option::is_none")] pub billing_mode: Option<String>, /// <p><p>One or more global secondary indexes (the maximum is 20) to be created on the table. Each global secondary index in the array includes the following:</p> <ul> <li> <p> <code>IndexName</code> - The name of the global secondary index. Must be unique only for this table.</p> <p/> </li> <li> <p> <code>KeySchema</code> - Specifies the key schema for the global secondary index.</p> </li> <li> <p> <code>Projection</code> - Specifies attributes that are copied (projected) from the table into the index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. Each attribute specification is composed of:</p> <ul> <li> <p> <code>ProjectionType</code> - One of the following:</p> <ul> <li> <p> <code>KEYS_ONLY</code> - Only the index and primary keys are projected into the index.</p> </li> <li> <p> <code>INCLUDE</code> - Only the specified table attributes are projected into the index. The list of projected attributes are in <code>NonKeyAttributes</code>.</p> </li> <li> <p> <code>ALL</code> - All of the table attributes are projected into the index.</p> </li> </ul> </li> <li> <p> <code>NonKeyAttributes</code> - A list of one or more non-key attribute names that are projected into the secondary index. The total count of attributes provided in <code>NonKeyAttributes</code>, summed across all of the secondary indexes, must not exceed 100. If you project the same attribute into two different indexes, this counts as two distinct attributes when determining the total.</p> </li> </ul> </li> <li> <p> <code>ProvisionedThroughput</code> - The provisioned throughput settings for the global secondary index, consisting of read and write capacity units.</p> </li> </ul></p> #[serde(rename = "GlobalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub global_secondary_indexes: Option<Vec<GlobalSecondaryIndex>>, /// <p>Specifies the attributes that make up the primary key for a table or an index. The attributes in <code>KeySchema</code> must also be defined in the <code>AttributeDefinitions</code> array. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/DataModel.html">Data Model</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>Each <code>KeySchemaElement</code> in the array is composed of:</p> <ul> <li> <p> <code>AttributeName</code> - The name of this key attribute.</p> </li> <li> <p> <code>KeyType</code> - The role that the key attribute will assume:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note> <p>For a simple primary key (partition key), you must provide exactly one element with a <code>KeyType</code> of <code>HASH</code>.</p> <p>For a composite primary key (partition key and sort key), you must provide exactly two elements, in this order: The first element must have a <code>KeyType</code> of <code>HASH</code>, and the second element must have a <code>KeyType</code> of <code>RANGE</code>.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#WorkingWithTables.primary.key">Specifying the Primary Key</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "KeySchema")] pub key_schema: Vec<KeySchemaElement>, /// <p><p>One or more local secondary indexes (the maximum is 5) to be created on the table. Each index is scoped to a given partition key value. There is a 10 GB size limit per partition key value; otherwise, the size of a local secondary index is unconstrained.</p> <p>Each local secondary index in the array includes the following:</p> <ul> <li> <p> <code>IndexName</code> - The name of the local secondary index. Must be unique only for this table.</p> <p/> </li> <li> <p> <code>KeySchema</code> - Specifies the key schema for the local secondary index. The key schema must begin with the same partition key as the table.</p> </li> <li> <p> <code>Projection</code> - Specifies attributes that are copied (projected) from the table into the index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. Each attribute specification is composed of:</p> <ul> <li> <p> <code>ProjectionType</code> - One of the following:</p> <ul> <li> <p> <code>KEYS_ONLY</code> - Only the index and primary keys are projected into the index.</p> </li> <li> <p> <code>INCLUDE</code> - Only the specified table attributes are projected into the index. The list of projected attributes are in <code>NonKeyAttributes</code>.</p> </li> <li> <p> <code>ALL</code> - All of the table attributes are projected into the index.</p> </li> </ul> </li> <li> <p> <code>NonKeyAttributes</code> - A list of one or more non-key attribute names that are projected into the secondary index. The total count of attributes provided in <code>NonKeyAttributes</code>, summed across all of the secondary indexes, must not exceed 100. If you project the same attribute into two different indexes, this counts as two distinct attributes when determining the total.</p> </li> </ul> </li> </ul></p> #[serde(rename = "LocalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub local_secondary_indexes: Option<Vec<LocalSecondaryIndex>>, /// <p>Represents the provisioned throughput settings for a specified table or index. The settings can be modified using the <code>UpdateTable</code> operation.</p> <p> If you set BillingMode as <code>PROVISIONED</code>, you must specify this property. If you set BillingMode as <code>PAY_PER_REQUEST</code>, you cannot specify this property. </p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughput>, /// <p>Represents the settings used to enable server-side encryption.</p> #[serde(rename = "SSESpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_specification: Option<SSESpecification>, /// <p><p>The settings for DynamoDB Streams on the table. These settings consist of:</p> <ul> <li> <p> <code>StreamEnabled</code> - Indicates whether Streams is to be enabled (true) or disabled (false).</p> </li> <li> <p> <code>StreamViewType</code> - When an item in the table is modified, <code>StreamViewType</code> determines what information is written to the table's stream. Valid values for <code>StreamViewType</code> are:</p> <ul> <li> <p> <code>KEYS<em>ONLY</code> - Only the key attributes of the modified item are written to the stream.</p> </li> <li> <p> <code>NEW</em>IMAGE</code> - The entire item, as it appears after it was modified, is written to the stream.</p> </li> <li> <p> <code>OLD<em>IMAGE</code> - The entire item, as it appeared before it was modified, is written to the stream.</p> </li> <li> <p> <code>NEW</em>AND<em>OLD</em>IMAGES</code> - Both the new and the old item images of the item are written to the stream.</p> </li> </ul> </li> </ul></p> #[serde(rename = "StreamSpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_specification: Option<StreamSpecification>, /// <p>The name of the table to create.</p> #[serde(rename = "TableName")] pub table_name: String, /// <p>A list of key-value pairs to label the table. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a>.</p> #[serde(rename = "Tags")] #[serde(skip_serializing_if = "Option::is_none")] pub tags: Option<Vec<Tag>>, } /// <p>Represents the output of a <code>CreateTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct CreateTableOutput { /// <p>Represents the properties of the table.</p> #[serde(rename = "TableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub table_description: Option<TableDescription>, } /// <p>Represents a request to perform a <code>DeleteItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct Delete { /// <p>A condition that must be satisfied in order for a conditional delete to succeed.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>One or more substitution tokens for attribute names in an expression.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The primary key of the item to be deleted. Each element consists of an attribute name and a value for that attribute.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, /// <p>Use <code>ReturnValuesOnConditionCheckFailure</code> to get the item attributes if the <code>Delete</code> condition fails. For <code>ReturnValuesOnConditionCheckFailure</code>, the valid values are: NONE and ALL_OLD.</p> #[serde(rename = "ReturnValuesOnConditionCheckFailure")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values_on_condition_check_failure: Option<String>, /// <p>Name of the table in which the item to be deleted resides.</p> #[serde(rename = "TableName")] pub table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DeleteBackupInput { /// <p>The ARN associated with the backup.</p> #[serde(rename = "BackupArn")] pub backup_arn: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DeleteBackupOutput { /// <p>Contains the description of the backup created for the table.</p> #[serde(rename = "BackupDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_description: Option<BackupDescription>, } /// <p>Represents a global secondary index to be deleted from an existing table.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DeleteGlobalSecondaryIndexAction { /// <p>The name of the global secondary index to be deleted.</p> #[serde(rename = "IndexName")] pub index_name: String, } /// <p>Represents the input of a <code>DeleteItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DeleteItemInput { /// <p>A condition that must be satisfied in order for a conditional <code>DeleteItem</code> to succeed.</p> <p>An expression can contain any of the following:</p> <ul> <li> <p>Functions: <code>attribute_exists | attribute_not_exists | attribute_type | contains | begins_with | size</code> </p> <p>These function names are case-sensitive.</p> </li> <li> <p>Comparison operators: <code>= | <> | < | > | <= | >= | BETWEEN | IN </code> </p> </li> <li> <p> Logical operators: <code>AND | OR | NOT</code> </p> </li> </ul> <p>For more information on condition expressions, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ConditionalOperator.html">ConditionalOperator</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionalOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub conditional_operator: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.Expected.html">Expected</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Expected")] #[serde(skip_serializing_if = "Option::is_none")] pub expected: Option<::std::collections::HashMap<String, ExpectedAttributeValue>>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> <p>Use the <b>:</b> (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the <i>ProductStatus</i> attribute was one of the following: </p> <p> <code>Available | Backordered | Discontinued</code> </p> <p>You would first need to specify <code>ExpressionAttributeValues</code> as follows:</p> <p> <code>{ ":avail":{"S":"Available"}, ":back":{"S":"Backordered"}, ":disc":{"S":"Discontinued"} }</code> </p> <p>You could then use these values in an expression, such as this:</p> <p> <code>ProductStatus IN (:avail, :back, :disc)</code> </p> <p>For more information on expression attribute values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>A map of attribute names to <code>AttributeValue</code> objects, representing the primary key of the item to delete.</p> <p>For the primary key, you must provide all of the attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Determines whether item collection metrics are returned. If set to <code>SIZE</code>, the response includes statistics about item collections, if any, that were modified during the operation are returned in the response. If set to <code>NONE</code> (the default), no statistics are returned.</p> #[serde(rename = "ReturnItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub return_item_collection_metrics: Option<String>, /// <p><p>Use <code>ReturnValues</code> if you want to get the item attributes as they appeared before they were deleted. For <code>DeleteItem</code>, the valid values are:</p> <ul> <li> <p> <code>NONE</code> - If <code>ReturnValues</code> is not specified, or if its value is <code>NONE</code>, then nothing is returned. (This setting is the default for <code>ReturnValues</code>.)</p> </li> <li> <p> <code>ALL<em>OLD</code> - The content of the old item is returned.</p> </li> </ul> <note> <p>The <code>ReturnValues</code> parameter is used by several DynamoDB operations; however, <code>DeleteItem</code> does not recognize any values other than <code>NONE</code> or <code>ALL</em>OLD</code>.</p> </note></p> #[serde(rename = "ReturnValues")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values: Option<String>, /// <p>The name of the table from which to delete the item.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>DeleteItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DeleteItemOutput { /// <p>A map of attribute names to <code>AttributeValue</code> objects, representing the item as it appeared before the <code>DeleteItem</code> operation. This map appears in the response only if <code>ReturnValues</code> was specified as <code>ALL_OLD</code> in the request.</p> #[serde(rename = "Attributes")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The capacity units consumed by the <code>DeleteItem</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p><p>Information about item collections, if any, that were affected by the <code>DeleteItem</code> operation. <code>ItemCollectionMetrics</code> is only returned if the <code>ReturnItemCollectionMetrics</code> parameter was specified. If the table does not have any local secondary indexes, this information is not returned in the response.</p> <p>Each <code>ItemCollectionMetrics</code> element consists of:</p> <ul> <li> <p> <code>ItemCollectionKey</code> - The partition key value of the item collection. This is the same as the partition key value of the item itself.</p> </li> <li> <p> <code>SizeEstimateRangeGB</code> - An estimate of item collection size, in gigabytes. This value is a two-element array containing a lower bound and an upper bound for the estimate. The estimate includes the size of all the items in the table, plus the size of all attributes projected into all of the local secondary indexes on that table. Use this estimate to measure whether a local secondary index is approaching its size limit.</p> <p>The estimate is subject to change over time; therefore, do not rely on the precision or accuracy of the estimate.</p> </li> </ul></p> #[serde(rename = "ItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_metrics: Option<ItemCollectionMetrics>, } /// <p>Represents a replica to be removed.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DeleteReplicaAction { /// <p>The region of the replica to be removed.</p> #[serde(rename = "RegionName")] pub region_name: String, } /// <p>Represents a request to perform a <code>DeleteItem</code> operation on an item.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct DeleteRequest { /// <p>A map of attribute name to attribute values, representing the primary key of the item to delete. All of the table's primary key attributes must be specified, and their data types must match those of the table's key schema.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, } /// <p>Represents the input of a <code>DeleteTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DeleteTableInput { /// <p>The name of the table to delete.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>DeleteTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DeleteTableOutput { /// <p>Represents the properties of a table.</p> #[serde(rename = "TableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub table_description: Option<TableDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeBackupInput { /// <p>The ARN associated with the backup.</p> #[serde(rename = "BackupArn")] pub backup_arn: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeBackupOutput { /// <p>Contains the description of the backup created for the table.</p> #[serde(rename = "BackupDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_description: Option<BackupDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeContinuousBackupsInput { /// <p>Name of the table for which the customer wants to check the continuous backups and point in time recovery settings.</p> #[serde(rename = "TableName")] pub table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeContinuousBackupsOutput { /// <p>Represents the continuous backups and point in time recovery settings on the table.</p> #[serde(rename = "ContinuousBackupsDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub continuous_backups_description: Option<ContinuousBackupsDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeEndpointsRequest {} #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeEndpointsResponse { /// <p>List of endpoints.</p> #[serde(rename = "Endpoints")] pub endpoints: Vec<Endpoint>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeGlobalTableInput { /// <p>The name of the global table.</p> #[serde(rename = "GlobalTableName")] pub global_table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeGlobalTableOutput { /// <p>Contains the details of the global table.</p> #[serde(rename = "GlobalTableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_description: Option<GlobalTableDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeGlobalTableSettingsInput { /// <p>The name of the global table to describe.</p> #[serde(rename = "GlobalTableName")] pub global_table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeGlobalTableSettingsOutput { /// <p>The name of the global table.</p> #[serde(rename = "GlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_name: Option<String>, /// <p>The region specific settings for the global table.</p> #[serde(rename = "ReplicaSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_settings: Option<Vec<ReplicaSettingsDescription>>, } /// <p>Represents the input of a <code>DescribeLimits</code> operation. Has no content.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeLimitsInput {} /// <p>Represents the output of a <code>DescribeLimits</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeLimitsOutput { /// <p>The maximum total read capacity units that your account allows you to provision across all of your tables in this region.</p> #[serde(rename = "AccountMaxReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub account_max_read_capacity_units: Option<i64>, /// <p>The maximum total write capacity units that your account allows you to provision across all of your tables in this region.</p> #[serde(rename = "AccountMaxWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub account_max_write_capacity_units: Option<i64>, /// <p>The maximum read capacity units that your account allows you to provision for a new table that you are creating in this region, including the read capacity units provisioned for its global secondary indexes (GSIs).</p> #[serde(rename = "TableMaxReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub table_max_read_capacity_units: Option<i64>, /// <p>The maximum write capacity units that your account allows you to provision for a new table that you are creating in this region, including the write capacity units provisioned for its global secondary indexes (GSIs).</p> #[serde(rename = "TableMaxWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub table_max_write_capacity_units: Option<i64>, } /// <p>Represents the input of a <code>DescribeTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeTableInput { /// <p>The name of the table to describe.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>DescribeTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeTableOutput { /// <p>The properties of the table.</p> #[serde(rename = "Table")] #[serde(skip_serializing_if = "Option::is_none")] pub table: Option<TableDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct DescribeTimeToLiveInput { /// <p>The name of the table to be described.</p> #[serde(rename = "TableName")] pub table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct DescribeTimeToLiveOutput { /// <p><p/></p> #[serde(rename = "TimeToLiveDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub time_to_live_description: Option<TimeToLiveDescription>, } /// <p>An endpoint information details.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct Endpoint { /// <p>IP address of the endpoint.</p> #[serde(rename = "Address")] pub address: String, /// <p>Endpoint cache time to live (TTL) value.</p> #[serde(rename = "CachePeriodInMinutes")] pub cache_period_in_minutes: i64, } /// <p>Represents a condition to be compared with an attribute value. This condition can be used with <code>DeleteItem</code>, <code>PutItem</code> or <code>UpdateItem</code> operations; if the comparison evaluates to true, the operation succeeds; if not, the operation fails. You can use <code>ExpectedAttributeValue</code> in one of two different ways:</p> <ul> <li> <p>Use <code>AttributeValueList</code> to specify one or more values to compare against an attribute. Use <code>ComparisonOperator</code> to specify how you want to perform the comparison. If the comparison evaluates to true, then the conditional operation succeeds.</p> </li> <li> <p>Use <code>Value</code> to specify a value that DynamoDB will compare against an attribute. If the values match, then <code>ExpectedAttributeValue</code> evaluates to true and the conditional operation succeeds. Optionally, you can also set <code>Exists</code> to false, indicating that you <i>do not</i> expect to find the attribute value in the table. In this case, the conditional operation succeeds only if the comparison evaluates to false.</p> </li> </ul> <p> <code>Value</code> and <code>Exists</code> are incompatible with <code>AttributeValueList</code> and <code>ComparisonOperator</code>. Note that if you use both sets of parameters at once, DynamoDB will return a <code>ValidationException</code> exception.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ExpectedAttributeValue { /// <p>One or more values to evaluate against the supplied attribute. The number of values in the list depends on the <code>ComparisonOperator</code> being used.</p> <p>For type Number, value comparisons are numeric.</p> <p>String value comparisons for greater than, equals, or less than are based on ASCII character code values. For example, <code>a</code> is greater than <code>A</code>, and <code>a</code> is greater than <code>B</code>. For a list of code values, see <a href="http://en.wikipedia.org/wiki/ASCII#ASCII_printable_characters">http://en.wikipedia.org/wiki/ASCII#ASCII_printable_characters</a>.</p> <p>For Binary, DynamoDB treats each byte of the binary data as unsigned when it compares binary values.</p> <p>For information on specifying data types in JSON, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/DataFormat.html">JSON Data Format</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributeValueList")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_value_list: Option<Vec<AttributeValue>>, /// <p><p>A comparator for evaluating attributes in the <code>AttributeValueList</code>. For example, equals, greater than, less than, etc.</p> <p>The following comparison operators are available:</p> <p> <code>EQ | NE | LE | LT | GE | GT | NOT<em>NULL | NULL | CONTAINS | NOT</em>CONTAINS | BEGINS<em>WITH | IN | BETWEEN</code> </p> <p>The following are descriptions of each comparison operator.</p> <ul> <li> <p> <code>EQ</code> : Equal. <code>EQ</code> is supported for all data types, including lists and maps.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, Binary, String Set, Number Set, or Binary Set. If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not equal <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>NE</code> : Not equal. <code>NE</code> is supported for all data types, including lists and maps.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String, Number, Binary, String Set, Number Set, or Binary Set. If an item contains an <code>AttributeValue</code> of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not equal <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>LE</code> : Less than or equal. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>LT</code> : Less than. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>GE</code> : Greater than or equal. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>GT</code> : Greater than. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not equal <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code>.</p> <p/> </li> <li> <p> <code>NOT</em>NULL</code> : The attribute exists. <code>NOT<em>NULL</code> is supported for all data types, including lists and maps.</p> <note> <p>This operator tests for the existence of an attribute, not its data type. If the data type of attribute "<code>a</code>" is null, and you evaluate it using <code>NOT</em>NULL</code>, the result is a Boolean <code>true</code>. This result is because the attribute "<code>a</code>" exists; its data type is not relevant to the <code>NOT<em>NULL</code> comparison operator.</p> </note> </li> <li> <p> <code>NULL</code> : The attribute does not exist. <code>NULL</code> is supported for all data types, including lists and maps.</p> <note> <p>This operator tests for the nonexistence of an attribute, not its data type. If the data type of attribute "<code>a</code>" is null, and you evaluate it using <code>NULL</code>, the result is a Boolean <code>false</code>. This is because the attribute "<code>a</code>" exists; its data type is not relevant to the <code>NULL</code> comparison operator.</p> </note> </li> <li> <p> <code>CONTAINS</code> : Checks for a subsequence, or value in a set.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If the target attribute of the comparison is of type String, then the operator checks for a substring match. If the target attribute of the comparison is of type Binary, then the operator looks for a subsequence of the target that matches the input. If the target attribute of the comparison is a set ("<code>SS</code>", "<code>NS</code>", or "<code>BS</code>"), then the operator evaluates to true if it finds an exact match with any member of the set.</p> <p>CONTAINS is supported for lists: When evaluating "<code>a CONTAINS b</code>", "<code>a</code>" can be a list; however, "<code>b</code>" cannot be a set, a map, or a list.</p> </li> <li> <p> <code>NOT</em>CONTAINS</code> : Checks for absence of a subsequence, or absence of a value in a set.</p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> element of type String, Number, or Binary (not a set type). If the target attribute of the comparison is a String, then the operator checks for the absence of a substring match. If the target attribute of the comparison is Binary, then the operator checks for the absence of a subsequence of the target that matches the input. If the target attribute of the comparison is a set ("<code>SS</code>", "<code>NS</code>", or "<code>BS</code>"), then the operator evaluates to true if it <i>does not</i> find an exact match with any member of the set.</p> <p>NOT<em>CONTAINS is supported for lists: When evaluating "<code>a NOT CONTAINS b</code>", "<code>a</code>" can be a list; however, "<code>b</code>" cannot be a set, a map, or a list.</p> </li> <li> <p> <code>BEGINS</em>WITH</code> : Checks for a prefix. </p> <p> <code>AttributeValueList</code> can contain only one <code>AttributeValue</code> of type String or Binary (not a Number or a set type). The target attribute of the comparison must be of type String or Binary (not a Number or a set type).</p> <p/> </li> <li> <p> <code>IN</code> : Checks for matching elements in a list.</p> <p> <code>AttributeValueList</code> can contain one or more <code>AttributeValue</code> elements of type String, Number, or Binary. These attributes are compared against an existing attribute of an item. If any elements of the input are equal to the item attribute, the expression evaluates to true.</p> </li> <li> <p> <code>BETWEEN</code> : Greater than or equal to the first value, and less than or equal to the second value. </p> <p> <code>AttributeValueList</code> must contain two <code>AttributeValue</code> elements of the same type, either String, Number, or Binary (not a set type). A target attribute matches if the target value is greater than, or equal to, the first element and less than, or equal to, the second element. If an item contains an <code>AttributeValue</code> element of a different type than the one provided in the request, the value does not match. For example, <code>{"S":"6"}</code> does not compare to <code>{"N":"6"}</code>. Also, <code>{"N":"6"}</code> does not compare to <code>{"NS":["6", "2", "1"]}</code> </p> </li> </ul></p> #[serde(rename = "ComparisonOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub comparison_operator: Option<String>, /// <p><p>Causes DynamoDB to evaluate the value before attempting a conditional operation:</p> <ul> <li> <p>If <code>Exists</code> is <code>true</code>, DynamoDB will check to see if that attribute value already exists in the table. If it is found, then the operation succeeds. If it is not found, the operation fails with a <code>ConditionCheckFailedException</code>.</p> </li> <li> <p>If <code>Exists</code> is <code>false</code>, DynamoDB assumes that the attribute value does not exist in the table. If in fact the value does not exist, then the assumption is valid and the operation succeeds. If the value is found, despite the assumption that it does not exist, the operation fails with a <code>ConditionCheckFailedException</code>.</p> </li> </ul> <p>The default setting for <code>Exists</code> is <code>true</code>. If you supply a <code>Value</code> all by itself, DynamoDB assumes the attribute exists: You don't have to set <code>Exists</code> to <code>true</code>, because it is implied.</p> <p>DynamoDB returns a <code>ValidationException</code> if:</p> <ul> <li> <p> <code>Exists</code> is <code>true</code> but there is no <code>Value</code> to check. (You expect a value to exist, but don't specify what that value is.)</p> </li> <li> <p> <code>Exists</code> is <code>false</code> but you also provide a <code>Value</code>. (You cannot expect an attribute to have a value, while also expecting it not to exist.)</p> </li> </ul></p> #[serde(rename = "Exists")] #[serde(skip_serializing_if = "Option::is_none")] pub exists: Option<bool>, /// <p>Represents the data for the expected attribute.</p> <p>Each attribute value is described as a name-value pair. The name is the data type, and the value is the data itself.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/HowItWorks.NamingRulesDataTypes.html#HowItWorks.DataTypes">Data Types</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Value")] #[serde(skip_serializing_if = "Option::is_none")] pub value: Option<AttributeValue>, } /// <p>Specifies an item and related attribute values to retrieve in a <code>TransactGetItem</code> object.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct Get { /// <p>One or more substitution tokens for attribute names in the ProjectionExpression parameter.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>A map of attribute names to <code>AttributeValue</code> objects that specifies the primary key of the item to retrieve.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, /// <p>A string that identifies one or more attributes of the specified item to retrieve from the table. The attributes in the expression must be separated by commas. If no attribute names are specified, then all attributes of the specified item are returned. If any of the requested attributes are not found, they do not appear in the result.</p> #[serde(rename = "ProjectionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_expression: Option<String>, /// <p>The name of the table from which to retrieve the specified item.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the input of a <code>GetItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct GetItemInput { /// <p>This is a legacy parameter. Use <code>ProjectionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.AttributesToGet.html">AttributesToGet</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributesToGet")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes_to_get: Option<Vec<String>>, /// <p>Determines the read consistency model: If set to <code>true</code>, then the operation uses strongly consistent reads; otherwise, the operation uses eventually consistent reads.</p> #[serde(rename = "ConsistentRead")] #[serde(skip_serializing_if = "Option::is_none")] pub consistent_read: Option<bool>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>A map of attribute names to <code>AttributeValue</code> objects, representing the primary key of the item to retrieve.</p> <p>For the primary key, you must provide all of the attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, /// <p>A string that identifies one or more attributes to retrieve from the table. These attributes can include scalars, sets, or elements of a JSON document. The attributes in the expression must be separated by commas.</p> <p>If no attribute names are specified, then all attributes will be returned. If any of the requested attributes are not found, they will not appear in the result.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProjectionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_expression: Option<String>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>The name of the table containing the requested item.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>GetItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct GetItemOutput { /// <p>The capacity units consumed by the <code>GetItem</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p>A map of attribute names to <code>AttributeValue</code> objects, as specified by <code>ProjectionExpression</code>.</p> #[serde(rename = "Item")] #[serde(skip_serializing_if = "Option::is_none")] pub item: Option<::std::collections::HashMap<String, AttributeValue>>, } /// <p>Represents the properties of a global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct GlobalSecondaryIndex { /// <p>The name of the global secondary index. The name must be unique among all other indexes on this table.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p><p>The complete key schema for a global secondary index, which consists of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] pub key_schema: Vec<KeySchemaElement>, /// <p>Represents attributes that are copied (projected) from the table into the global secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] pub projection: Projection, /// <p>Represents the provisioned throughput settings for the specified global secondary index.</p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughput>, } /// <p>Represents the properties of a global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct GlobalSecondaryIndexDescription { /// <p><p>Indicates whether the index is currently backfilling. <i>Backfilling</i> is the process of reading items from the table and determining whether they can be added to the index. (Not all items will qualify: For example, a partition key cannot have any duplicate values.) If an item can be added to the index, DynamoDB will do so. After all items have been processed, the backfilling operation is complete and <code>Backfilling</code> is false.</p> <note> <p>For indexes that were created during a <code>CreateTable</code> operation, the <code>Backfilling</code> attribute does not appear in the <code>DescribeTable</code> output.</p> </note></p> #[serde(rename = "Backfilling")] #[serde(skip_serializing_if = "Option::is_none")] pub backfilling: Option<bool>, /// <p>The Amazon Resource Name (ARN) that uniquely identifies the index.</p> #[serde(rename = "IndexArn")] #[serde(skip_serializing_if = "Option::is_none")] pub index_arn: Option<String>, /// <p>The name of the global secondary index.</p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p>The total size of the specified index, in bytes. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "IndexSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub index_size_bytes: Option<i64>, /// <p><p>The current state of the global secondary index:</p> <ul> <li> <p> <code>CREATING</code> - The index is being created.</p> </li> <li> <p> <code>UPDATING</code> - The index is being updated.</p> </li> <li> <p> <code>DELETING</code> - The index is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The index is ready for use.</p> </li> </ul></p> #[serde(rename = "IndexStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub index_status: Option<String>, /// <p>The number of items in the specified index. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "ItemCount")] #[serde(skip_serializing_if = "Option::is_none")] pub item_count: Option<i64>, /// <p><p>The complete key schema for a global secondary index, which consists of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] #[serde(skip_serializing_if = "Option::is_none")] pub key_schema: Option<Vec<KeySchemaElement>>, /// <p>Represents attributes that are copied (projected) from the table into the global secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] #[serde(skip_serializing_if = "Option::is_none")] pub projection: Option<Projection>, /// <p>Represents the provisioned throughput settings for the specified global secondary index.</p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughputDescription>, } /// <p>Represents the properties of a global secondary index for the table when the backup was created.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct GlobalSecondaryIndexInfo { /// <p>The name of the global secondary index.</p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p><p>The complete key schema for a global secondary index, which consists of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] #[serde(skip_serializing_if = "Option::is_none")] pub key_schema: Option<Vec<KeySchemaElement>>, /// <p>Represents attributes that are copied (projected) from the table into the global secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] #[serde(skip_serializing_if = "Option::is_none")] pub projection: Option<Projection>, /// <p>Represents the provisioned throughput settings for the specified global secondary index. </p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughput>, } /// <p><p>Represents one of the following:</p> <ul> <li> <p>A new global secondary index to be added to an existing table.</p> </li> <li> <p>New provisioned throughput parameters for an existing global secondary index.</p> </li> <li> <p>An existing global secondary index to be removed from an existing table.</p> </li> </ul></p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct GlobalSecondaryIndexUpdate { /// <p><p>The parameters required for creating a global secondary index on an existing table:</p> <ul> <li> <p> <code>IndexName </code> </p> </li> <li> <p> <code>KeySchema </code> </p> </li> <li> <p> <code>AttributeDefinitions </code> </p> </li> <li> <p> <code>Projection </code> </p> </li> <li> <p> <code>ProvisionedThroughput </code> </p> </li> </ul></p> #[serde(rename = "Create")] #[serde(skip_serializing_if = "Option::is_none")] pub create: Option<CreateGlobalSecondaryIndexAction>, /// <p>The name of an existing global secondary index to be removed.</p> #[serde(rename = "Delete")] #[serde(skip_serializing_if = "Option::is_none")] pub delete: Option<DeleteGlobalSecondaryIndexAction>, /// <p>The name of an existing global secondary index, along with new provisioned throughput settings to be applied to that index.</p> #[serde(rename = "Update")] #[serde(skip_serializing_if = "Option::is_none")] pub update: Option<UpdateGlobalSecondaryIndexAction>, } /// <p>Represents the properties of a global table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct GlobalTable { /// <p>The global table name.</p> #[serde(rename = "GlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_name: Option<String>, /// <p>The regions where the global table has replicas.</p> #[serde(rename = "ReplicationGroup")] #[serde(skip_serializing_if = "Option::is_none")] pub replication_group: Option<Vec<Replica>>, } /// <p>Contains details about the global table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct GlobalTableDescription { /// <p>The creation time of the global table.</p> #[serde(rename = "CreationDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub creation_date_time: Option<f64>, /// <p>The unique identifier of the global table.</p> #[serde(rename = "GlobalTableArn")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_arn: Option<String>, /// <p>The global table name.</p> #[serde(rename = "GlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_name: Option<String>, /// <p><p>The current state of the global table:</p> <ul> <li> <p> <code>CREATING</code> - The global table is being created.</p> </li> <li> <p> <code>UPDATING</code> - The global table is being updated.</p> </li> <li> <p> <code>DELETING</code> - The global table is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The global table is ready for use.</p> </li> </ul></p> #[serde(rename = "GlobalTableStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_status: Option<String>, /// <p>The regions where the global table has replicas.</p> #[serde(rename = "ReplicationGroup")] #[serde(skip_serializing_if = "Option::is_none")] pub replication_group: Option<Vec<ReplicaDescription>>, } /// <p>Represents the settings of a global secondary index for a global table that will be modified.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct GlobalTableGlobalSecondaryIndexSettingsUpdate { /// <p>The name of the global secondary index. The name must be unique among all other indexes on this table.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p>AutoScaling settings for managing a global secondary index's write capacity units.</p> #[serde(rename = "ProvisionedWriteCapacityAutoScalingSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_write_capacity_auto_scaling_settings_update: Option<AutoScalingSettingsUpdate>, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException.</code> </p> #[serde(rename = "ProvisionedWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_write_capacity_units: Option<i64>, } /// <p>Information about item collections, if any, that were affected by the operation. <code>ItemCollectionMetrics</code> is only returned if the request asked for it. If the table does not have any local secondary indexes, this information is not returned in the response.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ItemCollectionMetrics { /// <p>The partition key value of the item collection. This value is the same as the partition key value of the item.</p> #[serde(rename = "ItemCollectionKey")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_key: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>An estimate of item collection size, in gigabytes. This value is a two-element array containing a lower bound and an upper bound for the estimate. The estimate includes the size of all the items in the table, plus the size of all attributes projected into all of the local secondary indexes on that table. Use this estimate to measure whether a local secondary index is approaching its size limit.</p> <p>The estimate is subject to change over time; therefore, do not rely on the precision or accuracy of the estimate.</p> #[serde(rename = "SizeEstimateRangeGB")] #[serde(skip_serializing_if = "Option::is_none")] pub size_estimate_range_gb: Option<Vec<f64>>, } /// <p>Details for the requested item.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ItemResponse { /// <p>Map of attribute data consisting of the data type and attribute value.</p> #[serde(rename = "Item")] #[serde(skip_serializing_if = "Option::is_none")] pub item: Option<::std::collections::HashMap<String, AttributeValue>>, } /// <p>Represents <i>a single element</i> of a key schema. A key schema specifies the attributes that make up the primary key of a table, or the key attributes of an index.</p> <p>A <code>KeySchemaElement</code> represents exactly one attribute of the primary key. For example, a simple primary key would be represented by one <code>KeySchemaElement</code> (for the partition key). A composite primary key would require one <code>KeySchemaElement</code> for the partition key, and another <code>KeySchemaElement</code> for the sort key.</p> <p>A <code>KeySchemaElement</code> must be a scalar, top-level attribute (not a nested attribute). The data type must be one of String, Number, or Binary. The attribute cannot be nested within a List or a Map.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct KeySchemaElement { /// <p>The name of a key attribute.</p> #[serde(rename = "AttributeName")] pub attribute_name: String, /// <p><p>The role that this key attribute will assume:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeyType")] pub key_type: String, } /// <p>Represents a set of primary keys and, for each key, the attributes to retrieve from the table.</p> <p>For each primary key, you must provide <i>all</i> of the key attributes. For example, with a simple primary key, you only need to provide the partition key. For a composite primary key, you must provide <i>both</i> the partition key and the sort key.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct KeysAndAttributes { /// <p>This is a legacy parameter. Use <code>ProjectionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.html">Legacy Conditional Parameters</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributesToGet")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes_to_get: Option<Vec<String>>, /// <p>The consistency of a read operation. If set to <code>true</code>, then a strongly consistent read is used; otherwise, an eventually consistent read is used.</p> #[serde(rename = "ConsistentRead")] #[serde(skip_serializing_if = "Option::is_none")] pub consistent_read: Option<bool>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>The primary key attribute values that define the items and the attributes associated with the items.</p> #[serde(rename = "Keys")] pub keys: Vec<::std::collections::HashMap<String, AttributeValue>>, /// <p>A string that identifies one or more attributes to retrieve from the table. These attributes can include scalars, sets, or elements of a JSON document. The attributes in the <code>ProjectionExpression</code> must be separated by commas.</p> <p>If no attribute names are specified, then all attributes will be returned. If any of the requested attributes are not found, they will not appear in the result.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProjectionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_expression: Option<String>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ListBackupsInput { /// <p><p>The backups from the table specified by <code>BackupType</code> are listed.</p> <p>Where <code>BackupType</code> can be:</p> <ul> <li> <p> <code>USER</code> - On-demand backup created by you.</p> </li> <li> <p> <code>SYSTEM</code> - On-demand backup automatically created by DynamoDB.</p> </li> <li> <p> <code>ALL</code> - All types of on-demand backups (USER and SYSTEM).</p> </li> </ul></p> #[serde(rename = "BackupType")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_type: Option<String>, /// <p> <code>LastEvaluatedBackupArn</code> is the ARN of the backup last evaluated when the current page of results was returned, inclusive of the current page of results. This value may be specified as the <code>ExclusiveStartBackupArn</code> of a new <code>ListBackups</code> operation in order to fetch the next page of results. </p> #[serde(rename = "ExclusiveStartBackupArn")] #[serde(skip_serializing_if = "Option::is_none")] pub exclusive_start_backup_arn: Option<String>, /// <p>Maximum number of backups to return at once.</p> #[serde(rename = "Limit")] #[serde(skip_serializing_if = "Option::is_none")] pub limit: Option<i64>, /// <p>The backups from the table specified by <code>TableName</code> are listed. </p> #[serde(rename = "TableName")] #[serde(skip_serializing_if = "Option::is_none")] pub table_name: Option<String>, /// <p>Only backups created after this time are listed. <code>TimeRangeLowerBound</code> is inclusive.</p> #[serde(rename = "TimeRangeLowerBound")] #[serde(skip_serializing_if = "Option::is_none")] pub time_range_lower_bound: Option<f64>, /// <p>Only backups created before this time are listed. <code>TimeRangeUpperBound</code> is exclusive. </p> #[serde(rename = "TimeRangeUpperBound")] #[serde(skip_serializing_if = "Option::is_none")] pub time_range_upper_bound: Option<f64>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ListBackupsOutput { /// <p>List of <code>BackupSummary</code> objects.</p> #[serde(rename = "BackupSummaries")] #[serde(skip_serializing_if = "Option::is_none")] pub backup_summaries: Option<Vec<BackupSummary>>, /// <p> The ARN of the backup last evaluated when the current page of results was returned, inclusive of the current page of results. This value may be specified as the <code>ExclusiveStartBackupArn</code> of a new <code>ListBackups</code> operation in order to fetch the next page of results. </p> <p> If <code>LastEvaluatedBackupArn</code> is empty, then the last page of results has been processed and there are no more results to be retrieved. </p> <p> If <code>LastEvaluatedBackupArn</code> is not empty, this may or may not indicate there is more data to be returned. All results are guaranteed to have been returned if and only if no value for <code>LastEvaluatedBackupArn</code> is returned. </p> #[serde(rename = "LastEvaluatedBackupArn")] #[serde(skip_serializing_if = "Option::is_none")] pub last_evaluated_backup_arn: Option<String>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ListGlobalTablesInput { /// <p>The first global table name that this operation will evaluate.</p> #[serde(rename = "ExclusiveStartGlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub exclusive_start_global_table_name: Option<String>, /// <p>The maximum number of table names to return.</p> #[serde(rename = "Limit")] #[serde(skip_serializing_if = "Option::is_none")] pub limit: Option<i64>, /// <p>Lists the global tables in a specific region.</p> #[serde(rename = "RegionName")] #[serde(skip_serializing_if = "Option::is_none")] pub region_name: Option<String>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ListGlobalTablesOutput { /// <p>List of global table names.</p> #[serde(rename = "GlobalTables")] #[serde(skip_serializing_if = "Option::is_none")] pub global_tables: Option<Vec<GlobalTable>>, /// <p>Last evaluated global table name.</p> #[serde(rename = "LastEvaluatedGlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub last_evaluated_global_table_name: Option<String>, } /// <p>Represents the input of a <code>ListTables</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ListTablesInput { /// <p>The first table name that this operation will evaluate. Use the value that was returned for <code>LastEvaluatedTableName</code> in a previous operation, so that you can obtain the next page of results.</p> #[serde(rename = "ExclusiveStartTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub exclusive_start_table_name: Option<String>, /// <p>A maximum number of table names to return. If this parameter is not specified, the limit is 100.</p> #[serde(rename = "Limit")] #[serde(skip_serializing_if = "Option::is_none")] pub limit: Option<i64>, } /// <p>Represents the output of a <code>ListTables</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ListTablesOutput { /// <p>The name of the last table in the current page of results. Use this value as the <code>ExclusiveStartTableName</code> in a new request to obtain the next page of results, until all the table names are returned.</p> <p>If you do not receive a <code>LastEvaluatedTableName</code> value in the response, this means that there are no more table names to be retrieved.</p> #[serde(rename = "LastEvaluatedTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub last_evaluated_table_name: Option<String>, /// <p>The names of the tables associated with the current account at the current endpoint. The maximum size of this array is 100.</p> <p>If <code>LastEvaluatedTableName</code> also appears in the output, you can use this value as the <code>ExclusiveStartTableName</code> parameter in a subsequent <code>ListTables</code> request and obtain the next page of results.</p> #[serde(rename = "TableNames")] #[serde(skip_serializing_if = "Option::is_none")] pub table_names: Option<Vec<String>>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ListTagsOfResourceInput { /// <p>An optional string that, if supplied, must be copied from the output of a previous call to ListTagOfResource. When provided in this manner, this API fetches the next page of results.</p> #[serde(rename = "NextToken")] #[serde(skip_serializing_if = "Option::is_none")] pub next_token: Option<String>, /// <p>The Amazon DynamoDB resource with tags to be listed. This value is an Amazon Resource Name (ARN).</p> #[serde(rename = "ResourceArn")] pub resource_arn: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ListTagsOfResourceOutput { /// <p>If this value is returned, there are additional results to be displayed. To retrieve them, call ListTagsOfResource again, with NextToken set to this value.</p> #[serde(rename = "NextToken")] #[serde(skip_serializing_if = "Option::is_none")] pub next_token: Option<String>, /// <p>The tags currently associated with the Amazon DynamoDB resource.</p> #[serde(rename = "Tags")] #[serde(skip_serializing_if = "Option::is_none")] pub tags: Option<Vec<Tag>>, } /// <p>Represents the properties of a local secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct LocalSecondaryIndex { /// <p>The name of the local secondary index. The name must be unique among all other indexes on this table.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p><p>The complete key schema for the local secondary index, consisting of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] pub key_schema: Vec<KeySchemaElement>, /// <p>Represents attributes that are copied (projected) from the table into the local secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] pub projection: Projection, } /// <p>Represents the properties of a local secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct LocalSecondaryIndexDescription { /// <p>The Amazon Resource Name (ARN) that uniquely identifies the index.</p> #[serde(rename = "IndexArn")] #[serde(skip_serializing_if = "Option::is_none")] pub index_arn: Option<String>, /// <p>Represents the name of the local secondary index.</p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p>The total size of the specified index, in bytes. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "IndexSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub index_size_bytes: Option<i64>, /// <p>The number of items in the specified index. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "ItemCount")] #[serde(skip_serializing_if = "Option::is_none")] pub item_count: Option<i64>, /// <p><p>The complete key schema for the local secondary index, consisting of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] #[serde(skip_serializing_if = "Option::is_none")] pub key_schema: Option<Vec<KeySchemaElement>>, /// <p>Represents attributes that are copied (projected) from the table into the global secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] #[serde(skip_serializing_if = "Option::is_none")] pub projection: Option<Projection>, } /// <p>Represents the properties of a local secondary index for the table when the backup was created.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct LocalSecondaryIndexInfo { /// <p>Represents the name of the local secondary index.</p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p><p>The complete key schema for a local secondary index, which consists of one or more pairs of attribute names and key types:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note></p> #[serde(rename = "KeySchema")] #[serde(skip_serializing_if = "Option::is_none")] pub key_schema: Option<Vec<KeySchemaElement>>, /// <p>Represents attributes that are copied (projected) from the table into the global secondary index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. </p> #[serde(rename = "Projection")] #[serde(skip_serializing_if = "Option::is_none")] pub projection: Option<Projection>, } /// <p>The description of the point in time settings applied to the table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct PointInTimeRecoveryDescription { /// <p>Specifies the earliest point in time you can restore your table to. It You can restore your table to any point in time during the last 35 days. </p> #[serde(rename = "EarliestRestorableDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub earliest_restorable_date_time: Option<f64>, /// <p> <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. </p> #[serde(rename = "LatestRestorableDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub latest_restorable_date_time: Option<f64>, /// <p><p>The current state of point in time recovery:</p> <ul> <li> <p> <code>ENABLING</code> - Point in time recovery is being enabled.</p> </li> <li> <p> <code>ENABLED</code> - Point in time recovery is enabled.</p> </li> <li> <p> <code>DISABLED</code> - Point in time recovery is disabled.</p> </li> </ul></p> #[serde(rename = "PointInTimeRecoveryStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub point_in_time_recovery_status: Option<String>, } /// <p>Represents the settings used to enable point in time recovery.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct PointInTimeRecoverySpecification { /// <p>Indicates whether point in time recovery is enabled (true) or disabled (false) on the table.</p> #[serde(rename = "PointInTimeRecoveryEnabled")] pub point_in_time_recovery_enabled: bool, } /// <p>Represents attributes that are copied (projected) from the table into an index. These are in addition to the primary key attributes and index key attributes, which are automatically projected.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct Projection { /// <p>Represents the non-key attribute names which will be projected into the index.</p> <p>For local secondary indexes, the total count of <code>NonKeyAttributes</code> summed across all of the local secondary indexes, must not exceed 20. If you project the same attribute into two different indexes, this counts as two distinct attributes when determining the total.</p> #[serde(rename = "NonKeyAttributes")] #[serde(skip_serializing_if = "Option::is_none")] pub non_key_attributes: Option<Vec<String>>, /// <p><p>The set of attributes that are projected into the index:</p> <ul> <li> <p> <code>KEYS_ONLY</code> - Only the index and primary keys are projected into the index.</p> </li> <li> <p> <code>INCLUDE</code> - Only the specified table attributes are projected into the index. The list of projected attributes are in <code>NonKeyAttributes</code>.</p> </li> <li> <p> <code>ALL</code> - All of the table attributes are projected into the index.</p> </li> </ul></p> #[serde(rename = "ProjectionType")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_type: Option<String>, } /// <p>Represents the provisioned throughput settings for a specified table or index. The settings can be modified using the <code>UpdateTable</code> operation.</p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct ProvisionedThroughput { /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#ProvisionedThroughput">Specifying Read and Write Requirements</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>If read/write capacity mode is <code>PAY_PER_REQUEST</code> the value is set to 0.</p> #[serde(rename = "ReadCapacityUnits")] pub read_capacity_units: i64, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException</code>. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#ProvisionedThroughput">Specifying Read and Write Requirements</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>If read/write capacity mode is <code>PAY_PER_REQUEST</code> the value is set to 0.</p> #[serde(rename = "WriteCapacityUnits")] pub write_capacity_units: i64, } /// <p>Represents the provisioned throughput settings for the table, consisting of read and write capacity units, along with data about increases and decreases.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ProvisionedThroughputDescription { /// <p>The date and time of the last provisioned throughput decrease for this table.</p> #[serde(rename = "LastDecreaseDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub last_decrease_date_time: Option<f64>, /// <p>The date and time of the last provisioned throughput increase for this table.</p> #[serde(rename = "LastIncreaseDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub last_increase_date_time: Option<f64>, /// <p>The number of provisioned throughput decreases for this table during this UTC calendar day. For current maximums on provisioned throughput decreases, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "NumberOfDecreasesToday")] #[serde(skip_serializing_if = "Option::is_none")] pub number_of_decreases_today: Option<i64>, /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>. Eventually consistent reads require less effort than strongly consistent reads, so a setting of 50 <code>ReadCapacityUnits</code> per second provides 100 eventually consistent <code>ReadCapacityUnits</code> per second.</p> #[serde(rename = "ReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub read_capacity_units: Option<i64>, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException</code>.</p> #[serde(rename = "WriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub write_capacity_units: Option<i64>, } /// <p>Represents a request to perform a <code>PutItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct Put { /// <p>A condition that must be satisfied in order for a conditional update to succeed.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>One or more substitution tokens for attribute names in an expression.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>A map of attribute name to attribute values, representing the primary key of the item to be written by <code>PutItem</code>. All of the table's primary key attributes must be specified, and their data types must match those of the table's key schema. If any attributes are present in the item that are part of an index key schema for the table, their types must match the index key schema. </p> #[serde(rename = "Item")] pub item: ::std::collections::HashMap<String, AttributeValue>, /// <p>Use <code>ReturnValuesOnConditionCheckFailure</code> to get the item attributes if the <code>Put</code> condition fails. For <code>ReturnValuesOnConditionCheckFailure</code>, the valid values are: NONE and ALL_OLD.</p> #[serde(rename = "ReturnValuesOnConditionCheckFailure")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values_on_condition_check_failure: Option<String>, /// <p>Name of the table in which to write the item.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the input of a <code>PutItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct PutItemInput { /// <p>A condition that must be satisfied in order for a conditional <code>PutItem</code> operation to succeed.</p> <p>An expression can contain any of the following:</p> <ul> <li> <p>Functions: <code>attribute_exists | attribute_not_exists | attribute_type | contains | begins_with | size</code> </p> <p>These function names are case-sensitive.</p> </li> <li> <p>Comparison operators: <code>= | <> | < | > | <= | >= | BETWEEN | IN </code> </p> </li> <li> <p> Logical operators: <code>AND | OR | NOT</code> </p> </li> </ul> <p>For more information on condition expressions, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ConditionalOperator.html">ConditionalOperator</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionalOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub conditional_operator: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.Expected.html">Expected</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Expected")] #[serde(skip_serializing_if = "Option::is_none")] pub expected: Option<::std::collections::HashMap<String, ExpectedAttributeValue>>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> <p>Use the <b>:</b> (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the <i>ProductStatus</i> attribute was one of the following: </p> <p> <code>Available | Backordered | Discontinued</code> </p> <p>You would first need to specify <code>ExpressionAttributeValues</code> as follows:</p> <p> <code>{ ":avail":{"S":"Available"}, ":back":{"S":"Backordered"}, ":disc":{"S":"Discontinued"} }</code> </p> <p>You could then use these values in an expression, such as this:</p> <p> <code>ProductStatus IN (:avail, :back, :disc)</code> </p> <p>For more information on expression attribute values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>A map of attribute name/value pairs, one for each attribute. Only the primary key attributes are required; you can optionally provide other attribute name-value pairs for the item.</p> <p>You must provide all of the attributes for the primary key. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide both values for both the partition key and the sort key.</p> <p>If you specify any attributes that are part of an index key, then the data types for those attributes must match those of the schema in the table's attribute definition.</p> <p>For more information about primary keys, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/DataModel.html#DataModelPrimaryKey">Primary Key</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>Each element in the <code>Item</code> map is an <code>AttributeValue</code> object.</p> #[serde(rename = "Item")] pub item: ::std::collections::HashMap<String, AttributeValue>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Determines whether item collection metrics are returned. If set to <code>SIZE</code>, the response includes statistics about item collections, if any, that were modified during the operation are returned in the response. If set to <code>NONE</code> (the default), no statistics are returned.</p> #[serde(rename = "ReturnItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub return_item_collection_metrics: Option<String>, /// <p><p>Use <code>ReturnValues</code> if you want to get the item attributes as they appeared before they were updated with the <code>PutItem</code> request. For <code>PutItem</code>, the valid values are:</p> <ul> <li> <p> <code>NONE</code> - If <code>ReturnValues</code> is not specified, or if its value is <code>NONE</code>, then nothing is returned. (This setting is the default for <code>ReturnValues</code>.)</p> </li> <li> <p> <code>ALL<em>OLD</code> - If <code>PutItem</code> overwrote an attribute name-value pair, then the content of the old item is returned.</p> </li> </ul> <note> <p>The <code>ReturnValues</code> parameter is used by several DynamoDB operations; however, <code>PutItem</code> does not recognize any values other than <code>NONE</code> or <code>ALL</em>OLD</code>.</p> </note></p> #[serde(rename = "ReturnValues")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values: Option<String>, /// <p>The name of the table to contain the item.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>PutItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct PutItemOutput { /// <p>The attribute values as they appeared before the <code>PutItem</code> operation, but only if <code>ReturnValues</code> is specified as <code>ALL_OLD</code> in the request. Each element consists of an attribute name and an attribute value.</p> #[serde(rename = "Attributes")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The capacity units consumed by the <code>PutItem</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p><p>Information about item collections, if any, that were affected by the <code>PutItem</code> operation. <code>ItemCollectionMetrics</code> is only returned if the <code>ReturnItemCollectionMetrics</code> parameter was specified. If the table does not have any local secondary indexes, this information is not returned in the response.</p> <p>Each <code>ItemCollectionMetrics</code> element consists of:</p> <ul> <li> <p> <code>ItemCollectionKey</code> - The partition key value of the item collection. This is the same as the partition key value of the item itself.</p> </li> <li> <p> <code>SizeEstimateRangeGB</code> - An estimate of item collection size, in gigabytes. This value is a two-element array containing a lower bound and an upper bound for the estimate. The estimate includes the size of all the items in the table, plus the size of all attributes projected into all of the local secondary indexes on that table. Use this estimate to measure whether a local secondary index is approaching its size limit.</p> <p>The estimate is subject to change over time; therefore, do not rely on the precision or accuracy of the estimate.</p> </li> </ul></p> #[serde(rename = "ItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_metrics: Option<ItemCollectionMetrics>, } /// <p>Represents a request to perform a <code>PutItem</code> operation on an item.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct PutRequest { /// <p>A map of attribute name to attribute values, representing the primary key of an item to be processed by <code>PutItem</code>. All of the table's primary key attributes must be specified, and their data types must match those of the table's key schema. If any attributes are present in the item which are part of an index key schema for the table, their types must match the index key schema.</p> #[serde(rename = "Item")] pub item: ::std::collections::HashMap<String, AttributeValue>, } /// <p>Represents the input of a <code>Query</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct QueryInput { /// <p>This is a legacy parameter. Use <code>ProjectionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.AttributesToGet.html">AttributesToGet</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributesToGet")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes_to_get: Option<Vec<String>>, /// <p>This is a legacy parameter. Use <code>FilterExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ConditionalOperator.html">ConditionalOperator</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionalOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub conditional_operator: Option<String>, /// <p>Determines the read consistency model: If set to <code>true</code>, then the operation uses strongly consistent reads; otherwise, the operation uses eventually consistent reads.</p> <p>Strongly consistent reads are not supported on global secondary indexes. If you query a global secondary index with <code>ConsistentRead</code> set to <code>true</code>, you will receive a <code>ValidationException</code>.</p> #[serde(rename = "ConsistentRead")] #[serde(skip_serializing_if = "Option::is_none")] pub consistent_read: Option<bool>, /// <p>The primary key of the first item that this operation will evaluate. Use the value that was returned for <code>LastEvaluatedKey</code> in the previous operation.</p> <p>The data type for <code>ExclusiveStartKey</code> must be String, Number or Binary. No set data types are allowed.</p> #[serde(rename = "ExclusiveStartKey")] #[serde(skip_serializing_if = "Option::is_none")] pub exclusive_start_key: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> <p>Use the <b>:</b> (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the <i>ProductStatus</i> attribute was one of the following: </p> <p> <code>Available | Backordered | Discontinued</code> </p> <p>You would first need to specify <code>ExpressionAttributeValues</code> as follows:</p> <p> <code>{ ":avail":{"S":"Available"}, ":back":{"S":"Backordered"}, ":disc":{"S":"Discontinued"} }</code> </p> <p>You could then use these values in an expression, such as this:</p> <p> <code>ProductStatus IN (:avail, :back, :disc)</code> </p> <p>For more information on expression attribute values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>A string that contains conditions that DynamoDB applies after the <code>Query</code> operation, but before the data is returned to you. Items that do not satisfy the <code>FilterExpression</code> criteria are not returned.</p> <p>A <code>FilterExpression</code> does not allow key attributes. You cannot define a filter expression based on a partition key or a sort key.</p> <note> <p>A <code>FilterExpression</code> is applied after the items have already been read; the process of filtering does not consume any additional read capacity units.</p> </note> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html#FilteringResults">Filter Expressions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "FilterExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub filter_expression: Option<String>, /// <p>The name of an index to query. This index can be any local secondary index or global secondary index on the table. Note that if you use the <code>IndexName</code> parameter, you must also provide <code>TableName.</code> </p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p>The condition that specifies the key value(s) for items to be retrieved by the <code>Query</code> action.</p> <p>The condition must perform an equality test on a single partition key value.</p> <p>The condition can optionally perform one of several comparison tests on a single sort key value. This allows <code>Query</code> to retrieve one item with a given partition key value and sort key value, or several items that have the same partition key value but different sort key values.</p> <p>The partition key equality test is required, and must be specified in the following format:</p> <p> <code>partitionKeyName</code> <i>=</i> <code>:partitionkeyval</code> </p> <p>If you also want to provide a condition for the sort key, it must be combined using <code>AND</code> with the condition for the sort key. Following is an example, using the <b>=</b> comparison operator for the sort key:</p> <p> <code>partitionKeyName</code> <code>=</code> <code>:partitionkeyval</code> <code>AND</code> <code>sortKeyName</code> <code>=</code> <code>:sortkeyval</code> </p> <p>Valid comparisons for the sort key condition are as follows:</p> <ul> <li> <p> <code>sortKeyName</code> <code>=</code> <code>:sortkeyval</code> - true if the sort key value is equal to <code>:sortkeyval</code>.</p> </li> <li> <p> <code>sortKeyName</code> <code><</code> <code>:sortkeyval</code> - true if the sort key value is less than <code>:sortkeyval</code>.</p> </li> <li> <p> <code>sortKeyName</code> <code><=</code> <code>:sortkeyval</code> - true if the sort key value is less than or equal to <code>:sortkeyval</code>.</p> </li> <li> <p> <code>sortKeyName</code> <code>></code> <code>:sortkeyval</code> - true if the sort key value is greater than <code>:sortkeyval</code>.</p> </li> <li> <p> <code>sortKeyName</code> <code>>= </code> <code>:sortkeyval</code> - true if the sort key value is greater than or equal to <code>:sortkeyval</code>.</p> </li> <li> <p> <code>sortKeyName</code> <code>BETWEEN</code> <code>:sortkeyval1</code> <code>AND</code> <code>:sortkeyval2</code> - true if the sort key value is greater than or equal to <code>:sortkeyval1</code>, and less than or equal to <code>:sortkeyval2</code>.</p> </li> <li> <p> <code>begins_with (</code> <code>sortKeyName</code>, <code>:sortkeyval</code> <code>)</code> - true if the sort key value begins with a particular operand. (You cannot use this function with a sort key that is of type Number.) Note that the function name <code>begins_with</code> is case-sensitive.</p> </li> </ul> <p>Use the <code>ExpressionAttributeValues</code> parameter to replace tokens such as <code>:partitionval</code> and <code>:sortval</code> with actual values at runtime.</p> <p>You can optionally use the <code>ExpressionAttributeNames</code> parameter to replace the names of the partition key and sort key with placeholder tokens. This option might be necessary if an attribute name conflicts with a DynamoDB reserved word. For example, the following <code>KeyConditionExpression</code> parameter causes an error because <i>Size</i> is a reserved word:</p> <ul> <li> <p> <code>Size = :myval</code> </p> </li> </ul> <p>To work around this, define a placeholder (such a <code>#S</code>) to represent the attribute name <i>Size</i>. <code>KeyConditionExpression</code> then is as follows:</p> <ul> <li> <p> <code>#S = :myval</code> </p> </li> </ul> <p>For a list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>For more information on <code>ExpressionAttributeNames</code> and <code>ExpressionAttributeValues</code>, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ExpressionPlaceholders.html">Using Placeholders for Attribute Names and Values</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "KeyConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub key_condition_expression: Option<String>, /// <p>This is a legacy parameter. Use <code>KeyConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.KeyConditions.html">KeyConditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "KeyConditions")] #[serde(skip_serializing_if = "Option::is_none")] pub key_conditions: Option<::std::collections::HashMap<String, Condition>>, /// <p>The maximum number of items to evaluate (not necessarily the number of matching items). If DynamoDB processes the number of items up to the limit while processing the results, it stops the operation and returns the matching values up to that point, and a key in <code>LastEvaluatedKey</code> to apply in a subsequent operation, so that you can pick up where you left off. Also, if the processed data set size exceeds 1 MB before DynamoDB reaches this limit, it stops the operation and returns the matching values up to the limit, and a key in <code>LastEvaluatedKey</code> to apply in a subsequent operation to continue the operation. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html">Query and Scan</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Limit")] #[serde(skip_serializing_if = "Option::is_none")] pub limit: Option<i64>, /// <p>A string that identifies one or more attributes to retrieve from the table. These attributes can include scalars, sets, or elements of a JSON document. The attributes in the expression must be separated by commas.</p> <p>If no attribute names are specified, then all attributes will be returned. If any of the requested attributes are not found, they will not appear in the result.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProjectionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_expression: Option<String>, /// <p>This is a legacy parameter. Use <code>FilterExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.QueryFilter.html">QueryFilter</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "QueryFilter")] #[serde(skip_serializing_if = "Option::is_none")] pub query_filter: Option<::std::collections::HashMap<String, Condition>>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Specifies the order for index traversal: If <code>true</code> (default), the traversal is performed in ascending order; if <code>false</code>, the traversal is performed in descending order. </p> <p>Items with the same partition key value are stored in sorted order by sort key. If the sort key data type is Number, the results are stored in numeric order. For type String, the results are stored in order of UTF-8 bytes. For type Binary, DynamoDB treats each byte of the binary data as unsigned.</p> <p>If <code>ScanIndexForward</code> is <code>true</code>, DynamoDB returns the results in the order in which they are stored (by sort key value). This is the default behavior. If <code>ScanIndexForward</code> is <code>false</code>, DynamoDB reads the results in reverse order by sort key value, and then returns the results to the client.</p> #[serde(rename = "ScanIndexForward")] #[serde(skip_serializing_if = "Option::is_none")] pub scan_index_forward: Option<bool>, /// <p><p>The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.</p> <ul> <li> <p> <code>ALL<em>ATTRIBUTES</code> - Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index DynamoDB will fetch the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.</p> </li> <li> <p> <code>ALL</em>PROJECTED<em>ATTRIBUTES</code> - Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying <code>ALL</em>ATTRIBUTES</code>.</p> </li> <li> <p> <code>COUNT</code> - Returns the number of matching items, rather than the matching items themselves.</p> </li> <li> <p> <code>SPECIFIC<em>ATTRIBUTES</code> - Returns only the attributes listed in <code>AttributesToGet</code>. This return value is equivalent to specifying <code>AttributesToGet</code> without specifying any value for <code>Select</code>.</p> <p>If you query or scan a local secondary index and request only attributes that are projected into that index, the operation will read only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB will fetch each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.</p> <p>If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.</p> </li> </ul> <p>If neither <code>Select</code> nor <code>AttributesToGet</code> are specified, DynamoDB defaults to <code>ALL</em>ATTRIBUTES</code> when accessing a table, and <code>ALL<em>PROJECTED</em>ATTRIBUTES</code> when accessing an index. You cannot use both <code>Select</code> and <code>AttributesToGet</code> together in a single request, unless the value for <code>Select</code> is <code>SPECIFIC<em>ATTRIBUTES</code>. (This usage is equivalent to specifying <code>AttributesToGet</code> without any value for <code>Select</code>.)</p> <note> <p>If you use the <code>ProjectionExpression</code> parameter, then the value for <code>Select</code> can only be <code>SPECIFIC</em>ATTRIBUTES</code>. Any other value for <code>Select</code> will return an error.</p> </note></p> #[serde(rename = "Select")] #[serde(skip_serializing_if = "Option::is_none")] pub select: Option<String>, /// <p>The name of the table containing the requested items.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of a <code>Query</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct QueryOutput { /// <p>The capacity units consumed by the <code>Query</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p>The number of items in the response.</p> <p>If you used a <code>QueryFilter</code> in the request, then <code>Count</code> is the number of items returned after the filter was applied, and <code>ScannedCount</code> is the number of matching items before the filter was applied.</p> <p>If you did not use a filter in the request, then <code>Count</code> and <code>ScannedCount</code> are the same.</p> #[serde(rename = "Count")] #[serde(skip_serializing_if = "Option::is_none")] pub count: Option<i64>, /// <p>An array of item attributes that match the query criteria. Each element in this array consists of an attribute name and the value for that attribute.</p> #[serde(rename = "Items")] #[serde(skip_serializing_if = "Option::is_none")] pub items: Option<Vec<::std::collections::HashMap<String, AttributeValue>>>, /// <p>The primary key of the item where the operation stopped, inclusive of the previous result set. Use this value to start a new operation, excluding this value in the new request.</p> <p>If <code>LastEvaluatedKey</code> is empty, then the "last page" of results has been processed and there is no more data to be retrieved.</p> <p>If <code>LastEvaluatedKey</code> is not empty, it does not necessarily mean that there is more data in the result set. The only way to know when you have reached the end of the result set is when <code>LastEvaluatedKey</code> is empty.</p> #[serde(rename = "LastEvaluatedKey")] #[serde(skip_serializing_if = "Option::is_none")] pub last_evaluated_key: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The number of items evaluated, before any <code>QueryFilter</code> is applied. A high <code>ScannedCount</code> value with few, or no, <code>Count</code> results indicates an inefficient <code>Query</code> operation. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html#Count">Count and ScannedCount</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>If you did not use a filter in the request, then <code>ScannedCount</code> is the same as <code>Count</code>.</p> #[serde(rename = "ScannedCount")] #[serde(skip_serializing_if = "Option::is_none")] pub scanned_count: Option<i64>, } /// <p>Represents the properties of a replica.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct Replica { /// <p>The region where the replica needs to be created.</p> #[serde(rename = "RegionName")] #[serde(skip_serializing_if = "Option::is_none")] pub region_name: Option<String>, } /// <p>Contains the details of the replica.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ReplicaDescription { /// <p>The name of the region.</p> #[serde(rename = "RegionName")] #[serde(skip_serializing_if = "Option::is_none")] pub region_name: Option<String>, } /// <p>Represents the properties of a global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ReplicaGlobalSecondaryIndexSettingsDescription { /// <p>The name of the global secondary index. The name must be unique among all other indexes on this table.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p><p> The current status of the global secondary index:</p> <ul> <li> <p> <code>CREATING</code> - The global secondary index is being created.</p> </li> <li> <p> <code>UPDATING</code> - The global secondary index is being updated.</p> </li> <li> <p> <code>DELETING</code> - The global secondary index is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The global secondary index is ready for use.</p> </li> </ul></p> #[serde(rename = "IndexStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub index_status: Option<String>, /// <p>Autoscaling settings for a global secondary index replica's read capacity units.</p> #[serde(rename = "ProvisionedReadCapacityAutoScalingSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_read_capacity_auto_scaling_settings: Option<AutoScalingSettingsDescription>, /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>.</p> #[serde(rename = "ProvisionedReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_read_capacity_units: Option<i64>, /// <p>AutoScaling settings for a global secondary index replica's write capacity units.</p> #[serde(rename = "ProvisionedWriteCapacityAutoScalingSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_write_capacity_auto_scaling_settings: Option<AutoScalingSettingsDescription>, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException</code>.</p> #[serde(rename = "ProvisionedWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_write_capacity_units: Option<i64>, } /// <p>Represents the settings of a global secondary index for a global table that will be modified.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ReplicaGlobalSecondaryIndexSettingsUpdate { /// <p>The name of the global secondary index. The name must be unique among all other indexes on this table.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p>Autoscaling settings for managing a global secondary index replica's read capacity units.</p> #[serde(rename = "ProvisionedReadCapacityAutoScalingSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_read_capacity_auto_scaling_settings_update: Option<AutoScalingSettingsUpdate>, /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>.</p> #[serde(rename = "ProvisionedReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_read_capacity_units: Option<i64>, } /// <p>Represents the properties of a replica.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ReplicaSettingsDescription { /// <p>The region name of the replica.</p> #[serde(rename = "RegionName")] pub region_name: String, /// <p>The read/write capacity mode of the replica.</p> #[serde(rename = "ReplicaBillingModeSummary")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_billing_mode_summary: Option<BillingModeSummary>, /// <p>Replica global secondary index settings for the global table.</p> #[serde(rename = "ReplicaGlobalSecondaryIndexSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_global_secondary_index_settings: Option<Vec<ReplicaGlobalSecondaryIndexSettingsDescription>>, /// <p>Autoscaling settings for a global table replica's read capacity units.</p> #[serde(rename = "ReplicaProvisionedReadCapacityAutoScalingSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_read_capacity_auto_scaling_settings: Option<AutoScalingSettingsDescription>, /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#ProvisionedThroughput">Specifying Read and Write Requirements</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> #[serde(rename = "ReplicaProvisionedReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_read_capacity_units: Option<i64>, /// <p>AutoScaling settings for a global table replica's write capacity units.</p> #[serde(rename = "ReplicaProvisionedWriteCapacityAutoScalingSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_write_capacity_auto_scaling_settings: Option<AutoScalingSettingsDescription>, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException</code>. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#ProvisionedThroughput">Specifying Read and Write Requirements</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ReplicaProvisionedWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_write_capacity_units: Option<i64>, /// <p><p>The current state of the region:</p> <ul> <li> <p> <code>CREATING</code> - The region is being created.</p> </li> <li> <p> <code>UPDATING</code> - The region is being updated.</p> </li> <li> <p> <code>DELETING</code> - The region is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The region is ready for use.</p> </li> </ul></p> #[serde(rename = "ReplicaStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_status: Option<String>, } /// <p>Represents the settings for a global table in a region that will be modified.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ReplicaSettingsUpdate { /// <p>The region of the replica to be added.</p> #[serde(rename = "RegionName")] pub region_name: String, /// <p>Represents the settings of a global secondary index for a global table that will be modified.</p> #[serde(rename = "ReplicaGlobalSecondaryIndexSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_global_secondary_index_settings_update: Option<Vec<ReplicaGlobalSecondaryIndexSettingsUpdate>>, /// <p>Autoscaling settings for managing a global table replica's read capacity units.</p> #[serde(rename = "ReplicaProvisionedReadCapacityAutoScalingSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_read_capacity_auto_scaling_settings_update: Option<AutoScalingSettingsUpdate>, /// <p>The maximum number of strongly consistent reads consumed per second before DynamoDB returns a <code>ThrottlingException</code>. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#ProvisionedThroughput">Specifying Read and Write Requirements</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> #[serde(rename = "ReplicaProvisionedReadCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_provisioned_read_capacity_units: Option<i64>, } /// <p><p>Represents one of the following:</p> <ul> <li> <p>A new replica to be added to an existing global table.</p> </li> <li> <p>New parameters for an existing replica.</p> </li> <li> <p>An existing replica to be removed from an existing global table.</p> </li> </ul></p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ReplicaUpdate { /// <p>The parameters required for creating a replica on an existing global table.</p> #[serde(rename = "Create")] #[serde(skip_serializing_if = "Option::is_none")] pub create: Option<CreateReplicaAction>, /// <p>The name of the existing replica to be removed.</p> #[serde(rename = "Delete")] #[serde(skip_serializing_if = "Option::is_none")] pub delete: Option<DeleteReplicaAction>, } /// <p>Contains details for the restore.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct RestoreSummary { /// <p>Point in time or source backup time.</p> #[serde(rename = "RestoreDateTime")] pub restore_date_time: f64, /// <p>Indicates if a restore is in progress or not.</p> #[serde(rename = "RestoreInProgress")] pub restore_in_progress: bool, /// <p>ARN of the backup from which the table was restored.</p> #[serde(rename = "SourceBackupArn")] #[serde(skip_serializing_if = "Option::is_none")] pub source_backup_arn: Option<String>, /// <p>ARN of the source table of the backup that is being restored.</p> #[serde(rename = "SourceTableArn")] #[serde(skip_serializing_if = "Option::is_none")] pub source_table_arn: Option<String>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct RestoreTableFromBackupInput { /// <p>The ARN associated with the backup.</p> #[serde(rename = "BackupArn")] pub backup_arn: String, /// <p>The name of the new table to which the backup must be restored.</p> #[serde(rename = "TargetTableName")] pub target_table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct RestoreTableFromBackupOutput { /// <p>The description of the table created from an existing backup.</p> #[serde(rename = "TableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub table_description: Option<TableDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct RestoreTableToPointInTimeInput { /// <p>Time in the past to restore the table to.</p> #[serde(rename = "RestoreDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub restore_date_time: Option<f64>, /// <p>Name of the source table that is being restored.</p> #[serde(rename = "SourceTableName")] pub source_table_name: String, /// <p>The name of the new table to which it must be restored to.</p> #[serde(rename = "TargetTableName")] pub target_table_name: String, /// <p>Restore the table to the latest possible time. <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. </p> #[serde(rename = "UseLatestRestorableTime")] #[serde(skip_serializing_if = "Option::is_none")] pub use_latest_restorable_time: Option<bool>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct RestoreTableToPointInTimeOutput { /// <p>Represents the properties of a table.</p> #[serde(rename = "TableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub table_description: Option<TableDescription>, } /// <p>The description of the server-side encryption status on the specified table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct SSEDescription { /// <p>The KMS customer master key (CMK) ARN used for the KMS encryption.</p> #[serde(rename = "KMSMasterKeyArn")] #[serde(skip_serializing_if = "Option::is_none")] pub kms_master_key_arn: Option<String>, /// <p><p>Server-side encryption type. The only supported value is:</p> <ul> <li> <p> <code>KMS</code> - Server-side encryption which uses AWS Key Management Service. Key is stored in your account and is managed by AWS KMS (KMS charges apply).</p> </li> </ul></p> #[serde(rename = "SSEType")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_type: Option<String>, /// <p><p>Represents the current state of server-side encryption. The only supported values are:</p> <ul> <li> <p> <code>ENABLED</code> - Server-side encryption is enabled.</p> </li> <li> <p> <code>UPDATING</code> - Server-side encryption is being updated.</p> </li> </ul></p> #[serde(rename = "Status")] #[serde(skip_serializing_if = "Option::is_none")] pub status: Option<String>, } /// <p>Represents the settings used to enable server-side encryption.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct SSESpecification { /// <p>Indicates whether server-side encryption is done using an AWS managed CMK or an AWS owned CMK. If enabled (true), server-side encryption type is set to <code>KMS</code> and an AWS managed CMK is used (AWS KMS charges apply). If disabled (false) or not specified, server-side encryption is set to AWS owned CMK.</p> #[serde(rename = "Enabled")] #[serde(skip_serializing_if = "Option::is_none")] pub enabled: Option<bool>, /// <p>The KMS Customer Master Key (CMK) which should be used for the KMS encryption. To specify a CMK, use its key ID, Amazon Resource Name (ARN), alias name, or alias ARN. Note that you should only provide this parameter if the key is different from the default DynamoDB Customer Master Key alias/aws/dynamodb.</p> #[serde(rename = "KMSMasterKeyId")] #[serde(skip_serializing_if = "Option::is_none")] pub kms_master_key_id: Option<String>, /// <p><p>Server-side encryption type. The only supported value is:</p> <ul> <li> <p> <code>KMS</code> - Server-side encryption which uses AWS Key Management Service. Key is stored in your account and is managed by AWS KMS (KMS charges apply).</p> </li> </ul></p> #[serde(rename = "SSEType")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_type: Option<String>, } /// <p>Represents the input of a <code>Scan</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct ScanInput { /// <p>This is a legacy parameter. Use <code>ProjectionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.AttributesToGet.html">AttributesToGet</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributesToGet")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes_to_get: Option<Vec<String>>, /// <p>This is a legacy parameter. Use <code>FilterExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ConditionalOperator.html">ConditionalOperator</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionalOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub conditional_operator: Option<String>, /// <p>A Boolean value that determines the read consistency model during the scan:</p> <ul> <li> <p>If <code>ConsistentRead</code> is <code>false</code>, then the data returned from <code>Scan</code> might not contain the results from other recently completed write operations (PutItem, UpdateItem or DeleteItem).</p> </li> <li> <p>If <code>ConsistentRead</code> is <code>true</code>, then all of the write operations that completed before the <code>Scan</code> began are guaranteed to be contained in the <code>Scan</code> response.</p> </li> </ul> <p>The default setting for <code>ConsistentRead</code> is <code>false</code>.</p> <p>The <code>ConsistentRead</code> parameter is not supported on global secondary indexes. If you scan a global secondary index with <code>ConsistentRead</code> set to true, you will receive a <code>ValidationException</code>.</p> #[serde(rename = "ConsistentRead")] #[serde(skip_serializing_if = "Option::is_none")] pub consistent_read: Option<bool>, /// <p>The primary key of the first item that this operation will evaluate. Use the value that was returned for <code>LastEvaluatedKey</code> in the previous operation.</p> <p>The data type for <code>ExclusiveStartKey</code> must be String, Number or Binary. No set data types are allowed.</p> <p>In a parallel scan, a <code>Scan</code> request that includes <code>ExclusiveStartKey</code> must specify the same segment whose previous <code>Scan</code> returned the corresponding value of <code>LastEvaluatedKey</code>.</p> #[serde(rename = "ExclusiveStartKey")] #[serde(skip_serializing_if = "Option::is_none")] pub exclusive_start_key: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> <p>Use the <b>:</b> (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the <i>ProductStatus</i> attribute was one of the following: </p> <p> <code>Available | Backordered | Discontinued</code> </p> <p>You would first need to specify <code>ExpressionAttributeValues</code> as follows:</p> <p> <code>{ ":avail":{"S":"Available"}, ":back":{"S":"Backordered"}, ":disc":{"S":"Discontinued"} }</code> </p> <p>You could then use these values in an expression, such as this:</p> <p> <code>ProductStatus IN (:avail, :back, :disc)</code> </p> <p>For more information on expression attribute values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>A string that contains conditions that DynamoDB applies after the <code>Scan</code> operation, but before the data is returned to you. Items that do not satisfy the <code>FilterExpression</code> criteria are not returned.</p> <note> <p>A <code>FilterExpression</code> is applied after the items have already been read; the process of filtering does not consume any additional read capacity units.</p> </note> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html#FilteringResults">Filter Expressions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "FilterExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub filter_expression: Option<String>, /// <p>The name of a secondary index to scan. This index can be any local secondary index or global secondary index. Note that if you use the <code>IndexName</code> parameter, you must also provide <code>TableName</code>.</p> #[serde(rename = "IndexName")] #[serde(skip_serializing_if = "Option::is_none")] pub index_name: Option<String>, /// <p>The maximum number of items to evaluate (not necessarily the number of matching items). If DynamoDB processes the number of items up to the limit while processing the results, it stops the operation and returns the matching values up to that point, and a key in <code>LastEvaluatedKey</code> to apply in a subsequent operation, so that you can pick up where you left off. Also, if the processed data set size exceeds 1 MB before DynamoDB reaches this limit, it stops the operation and returns the matching values up to the limit, and a key in <code>LastEvaluatedKey</code> to apply in a subsequent operation to continue the operation. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html">Query and Scan</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Limit")] #[serde(skip_serializing_if = "Option::is_none")] pub limit: Option<i64>, /// <p>A string that identifies one or more attributes to retrieve from the specified table or index. These attributes can include scalars, sets, or elements of a JSON document. The attributes in the expression must be separated by commas.</p> <p>If no attribute names are specified, then all attributes will be returned. If any of the requested attributes are not found, they will not appear in the result.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProjectionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub projection_expression: Option<String>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>This is a legacy parameter. Use <code>FilterExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ScanFilter.html">ScanFilter</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ScanFilter")] #[serde(skip_serializing_if = "Option::is_none")] pub scan_filter: Option<::std::collections::HashMap<String, Condition>>, /// <p>For a parallel <code>Scan</code> request, <code>Segment</code> identifies an individual segment to be scanned by an application worker.</p> <p>Segment IDs are zero-based, so the first segment is always 0. For example, if you want to use four application threads to scan a table or an index, then the first thread specifies a <code>Segment</code> value of 0, the second thread specifies 1, and so on.</p> <p>The value of <code>LastEvaluatedKey</code> returned from a parallel <code>Scan</code> request must be used as <code>ExclusiveStartKey</code> with the same segment ID in a subsequent <code>Scan</code> operation.</p> <p>The value for <code>Segment</code> must be greater than or equal to 0, and less than the value provided for <code>TotalSegments</code>.</p> <p>If you provide <code>Segment</code>, you must also provide <code>TotalSegments</code>.</p> #[serde(rename = "Segment")] #[serde(skip_serializing_if = "Option::is_none")] pub segment: Option<i64>, /// <p><p>The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.</p> <ul> <li> <p> <code>ALL<em>ATTRIBUTES</code> - Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index DynamoDB will fetch the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.</p> </li> <li> <p> <code>ALL</em>PROJECTED<em>ATTRIBUTES</code> - Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying <code>ALL</em>ATTRIBUTES</code>.</p> </li> <li> <p> <code>COUNT</code> - Returns the number of matching items, rather than the matching items themselves.</p> </li> <li> <p> <code>SPECIFIC<em>ATTRIBUTES</code> - Returns only the attributes listed in <code>AttributesToGet</code>. This return value is equivalent to specifying <code>AttributesToGet</code> without specifying any value for <code>Select</code>.</p> <p>If you query or scan a local secondary index and request only attributes that are projected into that index, the operation will read only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB will fetch each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.</p> <p>If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.</p> </li> </ul> <p>If neither <code>Select</code> nor <code>AttributesToGet</code> are specified, DynamoDB defaults to <code>ALL</em>ATTRIBUTES</code> when accessing a table, and <code>ALL<em>PROJECTED</em>ATTRIBUTES</code> when accessing an index. You cannot use both <code>Select</code> and <code>AttributesToGet</code> together in a single request, unless the value for <code>Select</code> is <code>SPECIFIC<em>ATTRIBUTES</code>. (This usage is equivalent to specifying <code>AttributesToGet</code> without any value for <code>Select</code>.)</p> <note> <p>If you use the <code>ProjectionExpression</code> parameter, then the value for <code>Select</code> can only be <code>SPECIFIC</em>ATTRIBUTES</code>. Any other value for <code>Select</code> will return an error.</p> </note></p> #[serde(rename = "Select")] #[serde(skip_serializing_if = "Option::is_none")] pub select: Option<String>, /// <p>The name of the table containing the requested items; or, if you provide <code>IndexName</code>, the name of the table to which that index belongs.</p> #[serde(rename = "TableName")] pub table_name: String, /// <p>For a parallel <code>Scan</code> request, <code>TotalSegments</code> represents the total number of segments into which the <code>Scan</code> operation will be divided. The value of <code>TotalSegments</code> corresponds to the number of application workers that will perform the parallel scan. For example, if you want to use four application threads to scan a table or an index, specify a <code>TotalSegments</code> value of 4.</p> <p>The value for <code>TotalSegments</code> must be greater than or equal to 1, and less than or equal to 1000000. If you specify a <code>TotalSegments</code> value of 1, the <code>Scan</code> operation will be sequential rather than parallel.</p> <p>If you specify <code>TotalSegments</code>, you must also specify <code>Segment</code>.</p> #[serde(rename = "TotalSegments")] #[serde(skip_serializing_if = "Option::is_none")] pub total_segments: Option<i64>, } /// <p>Represents the output of a <code>Scan</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct ScanOutput { /// <p>The capacity units consumed by the <code>Scan</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p>The number of items in the response.</p> <p>If you set <code>ScanFilter</code> in the request, then <code>Count</code> is the number of items returned after the filter was applied, and <code>ScannedCount</code> is the number of matching items before the filter was applied.</p> <p>If you did not use a filter in the request, then <code>Count</code> is the same as <code>ScannedCount</code>.</p> #[serde(rename = "Count")] #[serde(skip_serializing_if = "Option::is_none")] pub count: Option<i64>, /// <p>An array of item attributes that match the scan criteria. Each element in this array consists of an attribute name and the value for that attribute.</p> #[serde(rename = "Items")] #[serde(skip_serializing_if = "Option::is_none")] pub items: Option<Vec<::std::collections::HashMap<String, AttributeValue>>>, /// <p>The primary key of the item where the operation stopped, inclusive of the previous result set. Use this value to start a new operation, excluding this value in the new request.</p> <p>If <code>LastEvaluatedKey</code> is empty, then the "last page" of results has been processed and there is no more data to be retrieved.</p> <p>If <code>LastEvaluatedKey</code> is not empty, it does not necessarily mean that there is more data in the result set. The only way to know when you have reached the end of the result set is when <code>LastEvaluatedKey</code> is empty.</p> #[serde(rename = "LastEvaluatedKey")] #[serde(skip_serializing_if = "Option::is_none")] pub last_evaluated_key: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The number of items evaluated, before any <code>ScanFilter</code> is applied. A high <code>ScannedCount</code> value with few, or no, <code>Count</code> results indicates an inefficient <code>Scan</code> operation. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/QueryAndScan.html#Count">Count and ScannedCount</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>If you did not use a filter in the request, then <code>ScannedCount</code> is the same as <code>Count</code>.</p> #[serde(rename = "ScannedCount")] #[serde(skip_serializing_if = "Option::is_none")] pub scanned_count: Option<i64>, } /// <p>Contains the details of the table when the backup was created. </p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct SourceTableDetails { /// <p><p>Controls how you are charged for read and write throughput and how you manage capacity. This setting can be changed later.</p> <ul> <li> <p> <code>PROVISIONED</code> - Sets the read/write capacity mode to <code>PROVISIONED</code>. We recommend using <code>PROVISIONED</code> for predictable workloads.</p> </li> <li> <p> <code>PAY<em>PER</em>REQUEST</code> - Sets the read/write capacity mode to <code>PAY<em>PER</em>REQUEST</code>. We recommend using <code>PAY<em>PER</em>REQUEST</code> for unpredictable workloads. </p> </li> </ul></p> #[serde(rename = "BillingMode")] #[serde(skip_serializing_if = "Option::is_none")] pub billing_mode: Option<String>, /// <p>Number of items in the table. Please note this is an approximate value. </p> #[serde(rename = "ItemCount")] #[serde(skip_serializing_if = "Option::is_none")] pub item_count: Option<i64>, /// <p>Schema of the table. </p> #[serde(rename = "KeySchema")] pub key_schema: Vec<KeySchemaElement>, /// <p>Read IOPs and Write IOPS on the table when the backup was created.</p> #[serde(rename = "ProvisionedThroughput")] pub provisioned_throughput: ProvisionedThroughput, /// <p>ARN of the table for which backup was created. </p> #[serde(rename = "TableArn")] #[serde(skip_serializing_if = "Option::is_none")] pub table_arn: Option<String>, /// <p>Time when the source table was created. </p> #[serde(rename = "TableCreationDateTime")] pub table_creation_date_time: f64, /// <p>Unique identifier for the table for which the backup was created. </p> #[serde(rename = "TableId")] pub table_id: String, /// <p>The name of the table for which the backup was created. </p> #[serde(rename = "TableName")] pub table_name: String, /// <p>Size of the table in bytes. Please note this is an approximate value.</p> #[serde(rename = "TableSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub table_size_bytes: Option<i64>, } /// <p>Contains the details of the features enabled on the table when the backup was created. For example, LSIs, GSIs, streams, TTL. </p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct SourceTableFeatureDetails { /// <p>Represents the GSI properties for the table when the backup was created. It includes the IndexName, KeySchema, Projection and ProvisionedThroughput for the GSIs on the table at the time of backup. </p> #[serde(rename = "GlobalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub global_secondary_indexes: Option<Vec<GlobalSecondaryIndexInfo>>, /// <p>Represents the LSI properties for the table when the backup was created. It includes the IndexName, KeySchema and Projection for the LSIs on the table at the time of backup. </p> #[serde(rename = "LocalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub local_secondary_indexes: Option<Vec<LocalSecondaryIndexInfo>>, /// <p>The description of the server-side encryption status on the table when the backup was created.</p> #[serde(rename = "SSEDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_description: Option<SSEDescription>, /// <p>Stream settings on the table when the backup was created.</p> #[serde(rename = "StreamDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_description: Option<StreamSpecification>, /// <p>Time to Live settings on the table when the backup was created.</p> #[serde(rename = "TimeToLiveDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub time_to_live_description: Option<TimeToLiveDescription>, } /// <p>Represents the DynamoDB Streams configuration for a table in DynamoDB.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct StreamSpecification { /// <p>Indicates whether DynamoDB Streams is enabled (true) or disabled (false) on the table.</p> #[serde(rename = "StreamEnabled")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_enabled: Option<bool>, /// <p><p> When an item in the table is modified, <code>StreamViewType</code> determines what information is written to the stream for this table. Valid values for <code>StreamViewType</code> are:</p> <ul> <li> <p> <code>KEYS<em>ONLY</code> - Only the key attributes of the modified item are written to the stream.</p> </li> <li> <p> <code>NEW</em>IMAGE</code> - The entire item, as it appears after it was modified, is written to the stream.</p> </li> <li> <p> <code>OLD<em>IMAGE</code> - The entire item, as it appeared before it was modified, is written to the stream.</p> </li> <li> <p> <code>NEW</em>AND<em>OLD</em>IMAGES</code> - Both the new and the old item images of the item are written to the stream.</p> </li> </ul></p> #[serde(rename = "StreamViewType")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_view_type: Option<String>, } /// <p>Represents the properties of a table.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct TableDescription { /// <p><p>An array of <code>AttributeDefinition</code> objects. Each of these objects describes one attribute in the table and index key schema.</p> <p>Each <code>AttributeDefinition</code> object in this array is composed of:</p> <ul> <li> <p> <code>AttributeName</code> - The name of the attribute.</p> </li> <li> <p> <code>AttributeType</code> - The data type for the attribute.</p> </li> </ul></p> #[serde(rename = "AttributeDefinitions")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_definitions: Option<Vec<AttributeDefinition>>, /// <p>Contains the details for the read/write capacity mode.</p> #[serde(rename = "BillingModeSummary")] #[serde(skip_serializing_if = "Option::is_none")] pub billing_mode_summary: Option<BillingModeSummary>, /// <p>The date and time when the table was created, in <a href="http://www.epochconverter.com/">UNIX epoch time</a> format.</p> #[serde(rename = "CreationDateTime")] #[serde(skip_serializing_if = "Option::is_none")] pub creation_date_time: Option<f64>, /// <p>The global secondary indexes, if any, on the table. Each index is scoped to a given partition key value. Each element is composed of:</p> <ul> <li> <p> <code>Backfilling</code> - If true, then the index is currently in the backfilling phase. Backfilling occurs only when a new global secondary index is added to the table; it is the process by which DynamoDB populates the new index with data from the table. (This attribute does not appear for indexes that were created during a <code>CreateTable</code> operation.)</p> </li> <li> <p> <code>IndexName</code> - The name of the global secondary index.</p> </li> <li> <p> <code>IndexSizeBytes</code> - The total size of the global secondary index, in bytes. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value. </p> </li> <li> <p> <code>IndexStatus</code> - The current status of the global secondary index:</p> <ul> <li> <p> <code>CREATING</code> - The index is being created.</p> </li> <li> <p> <code>UPDATING</code> - The index is being updated.</p> </li> <li> <p> <code>DELETING</code> - The index is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The index is ready for use.</p> </li> </ul> </li> <li> <p> <code>ItemCount</code> - The number of items in the global secondary index. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value. </p> </li> <li> <p> <code>KeySchema</code> - Specifies the complete index key schema. The attribute names in the key schema must be between 1 and 255 characters (inclusive). The key schema must begin with the same partition key as the table.</p> </li> <li> <p> <code>Projection</code> - Specifies attributes that are copied (projected) from the table into the index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. Each attribute specification is composed of:</p> <ul> <li> <p> <code>ProjectionType</code> - One of the following:</p> <ul> <li> <p> <code>KEYS_ONLY</code> - Only the index and primary keys are projected into the index.</p> </li> <li> <p> <code>INCLUDE</code> - Only the specified table attributes are projected into the index. The list of projected attributes are in <code>NonKeyAttributes</code>.</p> </li> <li> <p> <code>ALL</code> - All of the table attributes are projected into the index.</p> </li> </ul> </li> <li> <p> <code>NonKeyAttributes</code> - A list of one or more non-key attribute names that are projected into the secondary index. The total count of attributes provided in <code>NonKeyAttributes</code>, summed across all of the secondary indexes, must not exceed 20. If you project the same attribute into two different indexes, this counts as two distinct attributes when determining the total.</p> </li> </ul> </li> <li> <p> <code>ProvisionedThroughput</code> - The provisioned throughput settings for the global secondary index, consisting of read and write capacity units, along with data about increases and decreases. </p> </li> </ul> <p>If the table is in the <code>DELETING</code> state, no information about indexes will be returned.</p> #[serde(rename = "GlobalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub global_secondary_indexes: Option<Vec<GlobalSecondaryIndexDescription>>, /// <p>The number of items in the specified table. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "ItemCount")] #[serde(skip_serializing_if = "Option::is_none")] pub item_count: Option<i64>, /// <p>The primary key structure for the table. Each <code>KeySchemaElement</code> consists of:</p> <ul> <li> <p> <code>AttributeName</code> - The name of the attribute.</p> </li> <li> <p> <code>KeyType</code> - The role of the attribute:</p> <ul> <li> <p> <code>HASH</code> - partition key</p> </li> <li> <p> <code>RANGE</code> - sort key</p> </li> </ul> <note> <p>The partition key of an item is also known as its <i>hash attribute</i>. The term "hash attribute" derives from DynamoDB' usage of an internal hash function to evenly distribute data items across partitions, based on their partition key values.</p> <p>The sort key of an item is also known as its <i>range attribute</i>. The term "range attribute" derives from the way DynamoDB stores items with the same partition key physically close together, in sorted order by the sort key value.</p> </note> </li> </ul> <p>For more information about primary keys, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/DataModel.html#DataModelPrimaryKey">Primary Key</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "KeySchema")] #[serde(skip_serializing_if = "Option::is_none")] pub key_schema: Option<Vec<KeySchemaElement>>, /// <p>The Amazon Resource Name (ARN) that uniquely identifies the latest stream for this table.</p> #[serde(rename = "LatestStreamArn")] #[serde(skip_serializing_if = "Option::is_none")] pub latest_stream_arn: Option<String>, /// <p><p>A timestamp, in ISO 8601 format, for this stream.</p> <p>Note that <code>LatestStreamLabel</code> is not a unique identifier for the stream, because it is possible that a stream from another table might have the same timestamp. However, the combination of the following three elements is guaranteed to be unique:</p> <ul> <li> <p>the AWS customer ID.</p> </li> <li> <p>the table name.</p> </li> <li> <p>the <code>StreamLabel</code>.</p> </li> </ul></p> #[serde(rename = "LatestStreamLabel")] #[serde(skip_serializing_if = "Option::is_none")] pub latest_stream_label: Option<String>, /// <p>Represents one or more local secondary indexes on the table. Each index is scoped to a given partition key value. Tables with one or more local secondary indexes are subject to an item collection size limit, where the amount of data within a given item collection cannot exceed 10 GB. Each element is composed of:</p> <ul> <li> <p> <code>IndexName</code> - The name of the local secondary index.</p> </li> <li> <p> <code>KeySchema</code> - Specifies the complete index key schema. The attribute names in the key schema must be between 1 and 255 characters (inclusive). The key schema must begin with the same partition key as the table.</p> </li> <li> <p> <code>Projection</code> - Specifies attributes that are copied (projected) from the table into the index. These are in addition to the primary key attributes and index key attributes, which are automatically projected. Each attribute specification is composed of:</p> <ul> <li> <p> <code>ProjectionType</code> - One of the following:</p> <ul> <li> <p> <code>KEYS_ONLY</code> - Only the index and primary keys are projected into the index.</p> </li> <li> <p> <code>INCLUDE</code> - Only the specified table attributes are projected into the index. The list of projected attributes are in <code>NonKeyAttributes</code>.</p> </li> <li> <p> <code>ALL</code> - All of the table attributes are projected into the index.</p> </li> </ul> </li> <li> <p> <code>NonKeyAttributes</code> - A list of one or more non-key attribute names that are projected into the secondary index. The total count of attributes provided in <code>NonKeyAttributes</code>, summed across all of the secondary indexes, must not exceed 20. If you project the same attribute into two different indexes, this counts as two distinct attributes when determining the total.</p> </li> </ul> </li> <li> <p> <code>IndexSizeBytes</code> - Represents the total size of the index, in bytes. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> </li> <li> <p> <code>ItemCount</code> - Represents the number of items in the index. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> </li> </ul> <p>If the table is in the <code>DELETING</code> state, no information about indexes will be returned.</p> #[serde(rename = "LocalSecondaryIndexes")] #[serde(skip_serializing_if = "Option::is_none")] pub local_secondary_indexes: Option<Vec<LocalSecondaryIndexDescription>>, /// <p>The provisioned throughput settings for the table, consisting of read and write capacity units, along with data about increases and decreases.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughputDescription>, /// <p>Contains details for the restore.</p> #[serde(rename = "RestoreSummary")] #[serde(skip_serializing_if = "Option::is_none")] pub restore_summary: Option<RestoreSummary>, /// <p>The description of the server-side encryption status on the specified table.</p> #[serde(rename = "SSEDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_description: Option<SSEDescription>, /// <p>The current DynamoDB Streams configuration for the table.</p> #[serde(rename = "StreamSpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_specification: Option<StreamSpecification>, /// <p>The Amazon Resource Name (ARN) that uniquely identifies the table.</p> #[serde(rename = "TableArn")] #[serde(skip_serializing_if = "Option::is_none")] pub table_arn: Option<String>, /// <p>Unique identifier for the table for which the backup was created. </p> #[serde(rename = "TableId")] #[serde(skip_serializing_if = "Option::is_none")] pub table_id: Option<String>, /// <p>The name of the table.</p> #[serde(rename = "TableName")] #[serde(skip_serializing_if = "Option::is_none")] pub table_name: Option<String>, /// <p>The total size of the specified table, in bytes. DynamoDB updates this value approximately every six hours. Recent changes might not be reflected in this value.</p> #[serde(rename = "TableSizeBytes")] #[serde(skip_serializing_if = "Option::is_none")] pub table_size_bytes: Option<i64>, /// <p><p>The current state of the table:</p> <ul> <li> <p> <code>CREATING</code> - The table is being created.</p> </li> <li> <p> <code>UPDATING</code> - The table is being updated.</p> </li> <li> <p> <code>DELETING</code> - The table is being deleted.</p> </li> <li> <p> <code>ACTIVE</code> - The table is ready for use.</p> </li> </ul></p> #[serde(rename = "TableStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub table_status: Option<String>, } /// <p>Describes a tag. A tag is a key-value pair. You can add up to 50 tags to a single DynamoDB table. </p> <p> AWS-assigned tag names and values are automatically assigned the aws: prefix, which the user cannot assign. AWS-assigned tag names do not count towards the tag limit of 50. User-assigned tag names have the prefix user: in the Cost Allocation Report. You cannot backdate the application of a tag. </p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct Tag { /// <p>The key of the tag.Tag keys are case sensitive. Each DynamoDB table can only have up to one tag with the same key. If you try to add an existing tag (same key), the existing tag value will be updated to the new value. </p> #[serde(rename = "Key")] pub key: String, /// <p>The value of the tag. Tag values are case-sensitive and can be null.</p> #[serde(rename = "Value")] pub value: String, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct TagResourceInput { /// <p>Identifies the Amazon DynamoDB resource to which tags should be added. This value is an Amazon Resource Name (ARN).</p> #[serde(rename = "ResourceArn")] pub resource_arn: String, /// <p>The tags to be assigned to the Amazon DynamoDB resource.</p> #[serde(rename = "Tags")] pub tags: Vec<Tag>, } /// <p>The description of the Time to Live (TTL) status on the specified table. </p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct TimeToLiveDescription { /// <p> The name of the Time to Live attribute for items in the table.</p> #[serde(rename = "AttributeName")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_name: Option<String>, /// <p> The Time to Live status for the table.</p> #[serde(rename = "TimeToLiveStatus")] #[serde(skip_serializing_if = "Option::is_none")] pub time_to_live_status: Option<String>, } /// <p>Represents the settings used to enable or disable Time to Live for the specified table.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct TimeToLiveSpecification { /// <p>The name of the Time to Live attribute used to store the expiration time for items in the table.</p> #[serde(rename = "AttributeName")] pub attribute_name: String, /// <p>Indicates whether Time To Live is to be enabled (true) or disabled (false) on the table.</p> #[serde(rename = "Enabled")] pub enabled: bool, } /// <p>Specifies an item to be retrieved as part of the transaction.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct TransactGetItem { /// <p>Contains the primary key that identifies the item to get, together with the name of the table that contains the item, and optionally the specific attributes of the item to retrieve.</p> #[serde(rename = "Get")] pub get: Get, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct TransactGetItemsInput { /// <p>A value of <code>TOTAL</code> causes consumed capacity information to be returned, and a value of <code>NONE</code> prevents that information from being returned. No other value is valid.</p> #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>An ordered array of up to 10 <code>TransactGetItem</code> objects, each of which contains a <code>Get</code> structure.</p> #[serde(rename = "TransactItems")] pub transact_items: Vec<TransactGetItem>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct TransactGetItemsOutput { /// <p>If the <i>ReturnConsumedCapacity</i> value was <code>TOTAL</code>, this is an array of <code>ConsumedCapacity</code> objects, one for each table addressed by <code>TransactGetItem</code> objects in the <i>TransactItems</i> parameter. These <code>ConsumedCapacity</code> objects report the read-capacity units consumed by the <code>TransactGetItems</code> call in that table.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<Vec<ConsumedCapacity>>, /// <p>An ordered array of up to 10 <code>ItemResponse</code> objects, each of which corresponds to the <code>TransactGetItem</code> object in the same position in the <i>TransactItems</i> array. Each <code>ItemResponse</code> object contains a Map of the name-value pairs that are the projected attributes of the requested item.</p> <p>If a requested item could not be retrieved, the corresponding <code>ItemResponse</code> object is Null, or if the requested item has no projected attributes, the corresponding <code>ItemResponse</code> object is an empty Map. </p> #[serde(rename = "Responses")] #[serde(skip_serializing_if = "Option::is_none")] pub responses: Option<Vec<ItemResponse>>, } /// <p>A list of requests that can perform update, put, delete, or check operations on multiple items in one or more tables atomically.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct TransactWriteItem { /// <p>A request to perform a check item operation.</p> #[serde(rename = "ConditionCheck")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_check: Option<ConditionCheck>, /// <p>A request to perform a <code>DeleteItem</code> operation.</p> #[serde(rename = "Delete")] #[serde(skip_serializing_if = "Option::is_none")] pub delete: Option<Delete>, /// <p>A request to perform a <code>PutItem</code> operation.</p> #[serde(rename = "Put")] #[serde(skip_serializing_if = "Option::is_none")] pub put: Option<Put>, /// <p>A request to perform an <code>UpdateItem</code> operation.</p> #[serde(rename = "Update")] #[serde(skip_serializing_if = "Option::is_none")] pub update: Option<Update>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct TransactWriteItemsInput { /// <p>Providing a <code>ClientRequestToken</code> makes the call to <code>TransactWriteItems</code> idempotent, meaning that multiple identical calls have the same effect as one single call.</p> <p>Although multiple identical calls using the same client request token produce the same result on the server (no side effects), the responses to the calls may not be the same. If the <code>ReturnConsumedCapacity></code> parameter is set, then the initial <code>TransactWriteItems</code> call returns the amount of write capacity units consumed in making the changes, and subsequent <code>TransactWriteItems</code> calls with the same client token return the amount of read capacity units consumed in reading the item.</p> <p>A client request token is valid for 10 minutes after the first request that uses it completes. After 10 minutes, any request with the same client token is treated as a new request. Do not resubmit the same request with the same client token for more than 10 minutes or the result may not be idempotent.</p> <p>If you submit a request with the same client token but a change in other parameters within the 10 minute idempotency window, DynamoDB returns an <code>IdempotentParameterMismatch</code> exception.</p> #[serde(rename = "ClientRequestToken")] #[serde(skip_serializing_if = "Option::is_none")] pub client_request_token: Option<String>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Determines whether item collection metrics are returned. If set to <code>SIZE</code>, the response includes statistics about item collections (if any), that were modified during the operation and are returned in the response. If set to <code>NONE</code> (the default), no statistics are returned. </p> #[serde(rename = "ReturnItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub return_item_collection_metrics: Option<String>, /// <p>An ordered array of up to 10 <code>TransactWriteItem</code> objects, each of which contains a <code>ConditionCheck</code>, <code>Put</code>, <code>Update</code>, or <code>Delete</code> object. These can operate on items in different tables, but the tables must reside in the same AWS account and region, and no two of them can operate on the same item. </p> #[serde(rename = "TransactItems")] pub transact_items: Vec<TransactWriteItem>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct TransactWriteItemsOutput { /// <p>The capacity units consumed by the entire <code>TransactWriteItems</code> operation. The values of the list are ordered according to the ordering of the <code>TransactItems</code> request parameter. </p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<Vec<ConsumedCapacity>>, /// <p>A list of tables that were processed by <code>TransactWriteItems</code> and, for each table, information about any item collections that were affected by individual <code>UpdateItem</code>, <code>PutItem</code> or <code>DeleteItem</code> operations. </p> #[serde(rename = "ItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_metrics: Option<::std::collections::HashMap<String, Vec<ItemCollectionMetrics>>>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UntagResourceInput { /// <p>The Amazon DyanamoDB resource the tags will be removed from. This value is an Amazon Resource Name (ARN).</p> #[serde(rename = "ResourceArn")] pub resource_arn: String, /// <p>A list of tag keys. Existing tags of the resource whose keys are members of this list will be removed from the Amazon DynamoDB resource.</p> #[serde(rename = "TagKeys")] pub tag_keys: Vec<String>, } /// <p>Represents a request to perform an <code>UpdateItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct Update { /// <p>A condition that must be satisfied in order for a conditional update to succeed.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>One or more substitution tokens for attribute names in an expression.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The primary key of the item to be updated. Each element consists of an attribute name and a value for that attribute.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, /// <p>Use <code>ReturnValuesOnConditionCheckFailure</code> to get the item attributes if the <code>Update</code> condition fails. For <code>ReturnValuesOnConditionCheckFailure</code>, the valid values are: NONE, ALL_OLD, UPDATED_OLD, ALL_NEW, UPDATED_NEW.</p> #[serde(rename = "ReturnValuesOnConditionCheckFailure")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values_on_condition_check_failure: Option<String>, /// <p>Name of the table for the <code>UpdateItem</code> request.</p> #[serde(rename = "TableName")] pub table_name: String, /// <p>An expression that defines one or more attributes to be updated, the action to be performed on them, and new value(s) for them.</p> #[serde(rename = "UpdateExpression")] pub update_expression: String, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateContinuousBackupsInput { /// <p>Represents the settings used to enable point in time recovery.</p> #[serde(rename = "PointInTimeRecoverySpecification")] pub point_in_time_recovery_specification: PointInTimeRecoverySpecification, /// <p>The name of the table.</p> #[serde(rename = "TableName")] pub table_name: String, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateContinuousBackupsOutput { /// <p>Represents the continuous backups and point in time recovery settings on the table.</p> #[serde(rename = "ContinuousBackupsDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub continuous_backups_description: Option<ContinuousBackupsDescription>, } /// <p>Represents the new provisioned throughput settings to be applied to a global secondary index.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateGlobalSecondaryIndexAction { /// <p>The name of the global secondary index to be updated.</p> #[serde(rename = "IndexName")] pub index_name: String, /// <p>Represents the provisioned throughput settings for the specified global secondary index.</p> <p>For current minimum and maximum provisioned throughput values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ProvisionedThroughput")] pub provisioned_throughput: ProvisionedThroughput, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateGlobalTableInput { /// <p>The global table name.</p> #[serde(rename = "GlobalTableName")] pub global_table_name: String, /// <p>A list of regions that should be added or removed from the global table.</p> #[serde(rename = "ReplicaUpdates")] pub replica_updates: Vec<ReplicaUpdate>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateGlobalTableOutput { /// <p>Contains the details of the global table.</p> #[serde(rename = "GlobalTableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_description: Option<GlobalTableDescription>, } #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateGlobalTableSettingsInput { /// <p>The billing mode of the global table. If <code>GlobalTableBillingMode</code> is not specified, the global table defaults to <code>PROVISIONED</code> capacity billing mode.</p> #[serde(rename = "GlobalTableBillingMode")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_billing_mode: Option<String>, /// <p>Represents the settings of a global secondary index for a global table that will be modified.</p> #[serde(rename = "GlobalTableGlobalSecondaryIndexSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_global_secondary_index_settings_update: Option<Vec<GlobalTableGlobalSecondaryIndexSettingsUpdate>>, /// <p>The name of the global table</p> #[serde(rename = "GlobalTableName")] pub global_table_name: String, /// <p>AutoScaling settings for managing provisioned write capacity for the global table.</p> #[serde(rename = "GlobalTableProvisionedWriteCapacityAutoScalingSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_provisioned_write_capacity_auto_scaling_settings_update: Option<AutoScalingSettingsUpdate>, /// <p>The maximum number of writes consumed per second before DynamoDB returns a <code>ThrottlingException.</code> </p> #[serde(rename = "GlobalTableProvisionedWriteCapacityUnits")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_provisioned_write_capacity_units: Option<i64>, /// <p>Represents the settings for a global table in a region that will be modified.</p> #[serde(rename = "ReplicaSettingsUpdate")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_settings_update: Option<Vec<ReplicaSettingsUpdate>>, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateGlobalTableSettingsOutput { /// <p>The name of the global table.</p> #[serde(rename = "GlobalTableName")] #[serde(skip_serializing_if = "Option::is_none")] pub global_table_name: Option<String>, /// <p>The region specific settings for the global table.</p> #[serde(rename = "ReplicaSettings")] #[serde(skip_serializing_if = "Option::is_none")] pub replica_settings: Option<Vec<ReplicaSettingsDescription>>, } /// <p>Represents the input of an <code>UpdateItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateItemInput { /// <p>This is a legacy parameter. Use <code>UpdateExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.AttributeUpdates.html">AttributeUpdates</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "AttributeUpdates")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_updates: Option<::std::collections::HashMap<String, AttributeValueUpdate>>, /// <p>A condition that must be satisfied in order for a conditional update to succeed.</p> <p>An expression can contain any of the following:</p> <ul> <li> <p>Functions: <code>attribute_exists | attribute_not_exists | attribute_type | contains | begins_with | size</code> </p> <p>These function names are case-sensitive.</p> </li> <li> <p>Comparison operators: <code>= | <> | < | > | <= | >= | BETWEEN | IN </code> </p> </li> <li> <p> Logical operators: <code>AND | OR | NOT</code> </p> </li> </ul> <p>For more information on condition expressions, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub condition_expression: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.ConditionalOperator.html">ConditionalOperator</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConditionalOperator")] #[serde(skip_serializing_if = "Option::is_none")] pub conditional_operator: Option<String>, /// <p>This is a legacy parameter. Use <code>ConditionExpression</code> instead. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/LegacyConditionalParameters.Expected.html">Expected</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "Expected")] #[serde(skip_serializing_if = "Option::is_none")] pub expected: Option<::std::collections::HashMap<String, ExpectedAttributeValue>>, /// <p>One or more substitution tokens for attribute names in an expression. The following are some use cases for using <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p>To access an attribute whose name conflicts with a DynamoDB reserved word.</p> </li> <li> <p>To create a placeholder for repeating occurrences of an attribute name in an expression.</p> </li> <li> <p>To prevent special characters in an attribute name from being misinterpreted in an expression.</p> </li> </ul> <p>Use the <b>#</b> character in an expression to dereference an attribute name. For example, consider the following attribute name:</p> <ul> <li> <p> <code>Percentile</code> </p> </li> </ul> <p>The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ReservedWords.html">Reserved Words</a> in the <i>Amazon DynamoDB Developer Guide</i>). To work around this, you could specify the following for <code>ExpressionAttributeNames</code>:</p> <ul> <li> <p> <code>{"#P":"Percentile"}</code> </p> </li> </ul> <p>You could then use this substitution in an expression, as in this example:</p> <ul> <li> <p> <code>#P = :val</code> </p> </li> </ul> <note> <p>Tokens that begin with the <b>:</b> character are <i>expression attribute values</i>, which are placeholders for the actual value at runtime.</p> </note> <p>For more information on expression attribute names, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.AccessingItemAttributes.html">Accessing Item Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeNames")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_names: Option<::std::collections::HashMap<String, String>>, /// <p>One or more values that can be substituted in an expression.</p> <p>Use the <b>:</b> (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the <i>ProductStatus</i> attribute was one of the following: </p> <p> <code>Available | Backordered | Discontinued</code> </p> <p>You would first need to specify <code>ExpressionAttributeValues</code> as follows:</p> <p> <code>{ ":avail":{"S":"Available"}, ":back":{"S":"Backordered"}, ":disc":{"S":"Discontinued"} }</code> </p> <p>You could then use these values in an expression, such as this:</p> <p> <code>ProductStatus IN (:avail, :back, :disc)</code> </p> <p>For more information on expression attribute values, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.SpecifyingConditions.html">Specifying Conditions</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ExpressionAttributeValues")] #[serde(skip_serializing_if = "Option::is_none")] pub expression_attribute_values: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The primary key of the item to be updated. Each element consists of an attribute name and a value for that attribute.</p> <p>For the primary key, you must provide all of the attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.</p> #[serde(rename = "Key")] pub key: ::std::collections::HashMap<String, AttributeValue>, #[serde(rename = "ReturnConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub return_consumed_capacity: Option<String>, /// <p>Determines whether item collection metrics are returned. If set to <code>SIZE</code>, the response includes statistics about item collections, if any, that were modified during the operation are returned in the response. If set to <code>NONE</code> (the default), no statistics are returned.</p> #[serde(rename = "ReturnItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub return_item_collection_metrics: Option<String>, /// <p>Use <code>ReturnValues</code> if you want to get the item attributes as they appear before or after they are updated. For <code>UpdateItem</code>, the valid values are:</p> <ul> <li> <p> <code>NONE</code> - If <code>ReturnValues</code> is not specified, or if its value is <code>NONE</code>, then nothing is returned. (This setting is the default for <code>ReturnValues</code>.)</p> </li> <li> <p> <code>ALL_OLD</code> - Returns all of the attributes of the item, as they appeared before the UpdateItem operation.</p> </li> <li> <p> <code>UPDATED_OLD</code> - Returns only the updated attributes, as they appeared before the UpdateItem operation.</p> </li> <li> <p> <code>ALL_NEW</code> - Returns all of the attributes of the item, as they appear after the UpdateItem operation.</p> </li> <li> <p> <code>UPDATED_NEW</code> - Returns only the updated attributes, as they appear after the UpdateItem operation.</p> </li> </ul> <p>There is no additional cost associated with requesting a return value aside from the small network and processing overhead of receiving a larger response. No read capacity units are consumed.</p> <p>The values returned are strongly consistent.</p> #[serde(rename = "ReturnValues")] #[serde(skip_serializing_if = "Option::is_none")] pub return_values: Option<String>, /// <p>The name of the table containing the item to update.</p> #[serde(rename = "TableName")] pub table_name: String, /// <p>An expression that defines one or more attributes to be updated, the action to be performed on them, and new value(s) for them.</p> <p>The following action values are available for <code>UpdateExpression</code>.</p> <ul> <li> <p> <code>SET</code> - Adds one or more attributes and values to an item. If any of these attribute already exist, they are replaced by the new values. You can also use <code>SET</code> to add or subtract from an attribute that is of type Number. For example: <code>SET myNum = myNum + :val</code> </p> <p> <code>SET</code> supports the following functions:</p> <ul> <li> <p> <code>if_not_exists (path, operand)</code> - if the item does not contain an attribute at the specified path, then <code>if_not_exists</code> evaluates to operand; otherwise, it evaluates to path. You can use this function to avoid overwriting an attribute that may already be present in the item.</p> </li> <li> <p> <code>list_append (operand, operand)</code> - evaluates to a list with a new element added to it. You can append the new element to the start or the end of the list by reversing the order of the operands.</p> </li> </ul> <p>These function names are case-sensitive.</p> </li> <li> <p> <code>REMOVE</code> - Removes one or more attributes from an item.</p> </li> <li> <p> <code>ADD</code> - Adds the specified value to the item, if the attribute does not already exist. If the attribute does exist, then the behavior of <code>ADD</code> depends on the data type of the attribute:</p> <ul> <li> <p>If the existing attribute is a number, and if <code>Value</code> is also a number, then <code>Value</code> is mathematically added to the existing attribute. If <code>Value</code> is a negative number, then it is subtracted from the existing attribute.</p> <note> <p>If you use <code>ADD</code> to increment or decrement a number value for an item that doesn't exist before the update, DynamoDB uses <code>0</code> as the initial value.</p> <p>Similarly, if you use <code>ADD</code> for an existing item to increment or decrement an attribute value that doesn't exist before the update, DynamoDB uses <code>0</code> as the initial value. For example, suppose that the item you want to update doesn't have an attribute named <i>itemcount</i>, but you decide to <code>ADD</code> the number <code>3</code> to this attribute anyway. DynamoDB will create the <i>itemcount</i> attribute, set its initial value to <code>0</code>, and finally add <code>3</code> to it. The result will be a new <i>itemcount</i> attribute in the item, with a value of <code>3</code>.</p> </note> </li> <li> <p>If the existing data type is a set and if <code>Value</code> is also a set, then <code>Value</code> is added to the existing set. For example, if the attribute value is the set <code>[1,2]</code>, and the <code>ADD</code> action specified <code>[3]</code>, then the final attribute value is <code>[1,2,3]</code>. An error occurs if an <code>ADD</code> action is specified for a set attribute and the attribute type specified does not match the existing set type. </p> <p>Both sets must have the same primitive data type. For example, if the existing data type is a set of strings, the <code>Value</code> must also be a set of strings.</p> </li> </ul> <important> <p>The <code>ADD</code> action only supports Number and set data types. In addition, <code>ADD</code> can only be used on top-level attributes, not nested attributes.</p> </important> </li> <li> <p> <code>DELETE</code> - Deletes an element from a set.</p> <p>If a set of values is specified, then those values are subtracted from the old set. For example, if the attribute value was the set <code>[a,b,c]</code> and the <code>DELETE</code> action specifies <code>[a,c]</code>, then the final attribute value is <code>[b]</code>. Specifying an empty set is an error.</p> <important> <p>The <code>DELETE</code> action only supports set data types. In addition, <code>DELETE</code> can only be used on top-level attributes, not nested attributes.</p> </important> </li> </ul> <p>You can have many actions in a single expression, such as the following: <code>SET a=:value1, b=:value2 DELETE :value3, :value4, :value5</code> </p> <p>For more information on update expressions, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.Modifying.html">Modifying Items and Attributes</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "UpdateExpression")] #[serde(skip_serializing_if = "Option::is_none")] pub update_expression: Option<String>, } /// <p>Represents the output of an <code>UpdateItem</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateItemOutput { /// <p>A map of attribute values as they appear before or after the <code>UpdateItem</code> operation, as determined by the <code>ReturnValues</code> parameter.</p> <p>The <code>Attributes</code> map is only present if <code>ReturnValues</code> was specified as something other than <code>NONE</code> in the request. Each element represents one attribute.</p> #[serde(rename = "Attributes")] #[serde(skip_serializing_if = "Option::is_none")] pub attributes: Option<::std::collections::HashMap<String, AttributeValue>>, /// <p>The capacity units consumed by the <code>UpdateItem</code> operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. <code>ConsumedCapacity</code> is only returned if the <code>ReturnConsumedCapacity</code> parameter was specified. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html">Provisioned Throughput</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> #[serde(rename = "ConsumedCapacity")] #[serde(skip_serializing_if = "Option::is_none")] pub consumed_capacity: Option<ConsumedCapacity>, /// <p><p>Information about item collections, if any, that were affected by the <code>UpdateItem</code> operation. <code>ItemCollectionMetrics</code> is only returned if the <code>ReturnItemCollectionMetrics</code> parameter was specified. If the table does not have any local secondary indexes, this information is not returned in the response.</p> <p>Each <code>ItemCollectionMetrics</code> element consists of:</p> <ul> <li> <p> <code>ItemCollectionKey</code> - The partition key value of the item collection. This is the same as the partition key value of the item itself.</p> </li> <li> <p> <code>SizeEstimateRangeGB</code> - An estimate of item collection size, in gigabytes. This value is a two-element array containing a lower bound and an upper bound for the estimate. The estimate includes the size of all the items in the table, plus the size of all attributes projected into all of the local secondary indexes on that table. Use this estimate to measure whether a local secondary index is approaching its size limit.</p> <p>The estimate is subject to change over time; therefore, do not rely on the precision or accuracy of the estimate.</p> </li> </ul></p> #[serde(rename = "ItemCollectionMetrics")] #[serde(skip_serializing_if = "Option::is_none")] pub item_collection_metrics: Option<ItemCollectionMetrics>, } /// <p>Represents the input of an <code>UpdateTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateTableInput { /// <p>An array of attributes that describe the key schema for the table and indexes. If you are adding a new global secondary index to the table, <code>AttributeDefinitions</code> must include the key element(s) of the new index.</p> #[serde(rename = "AttributeDefinitions")] #[serde(skip_serializing_if = "Option::is_none")] pub attribute_definitions: Option<Vec<AttributeDefinition>>, /// <p><p>Controls how you are charged for read and write throughput and how you manage capacity. When switching from pay-per-request to provisioned capacity, initial provisioned capacity values must be set. The initial provisioned capacity values are estimated based on the consumed read and write capacity of your table and global secondary indexes over the past 30 minutes.</p> <ul> <li> <p> <code>PROVISIONED</code> - Sets the billing mode to <code>PROVISIONED</code>. We recommend using <code>PROVISIONED</code> for predictable workloads.</p> </li> <li> <p> <code>PAY<em>PER</em>REQUEST</code> - Sets the billing mode to <code>PAY<em>PER</em>REQUEST</code>. We recommend using <code>PAY<em>PER</em>REQUEST</code> for unpredictable workloads. </p> </li> </ul></p> #[serde(rename = "BillingMode")] #[serde(skip_serializing_if = "Option::is_none")] pub billing_mode: Option<String>, /// <p>An array of one or more global secondary indexes for the table. For each index in the array, you can request one action:</p> <ul> <li> <p> <code>Create</code> - add a new global secondary index to the table.</p> </li> <li> <p> <code>Update</code> - modify the provisioned throughput settings of an existing global secondary index.</p> </li> <li> <p> <code>Delete</code> - remove a global secondary index from the table.</p> </li> </ul> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/GSI.OnlineOps.html">Managing Global Secondary Indexes</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> #[serde(rename = "GlobalSecondaryIndexUpdates")] #[serde(skip_serializing_if = "Option::is_none")] pub global_secondary_index_updates: Option<Vec<GlobalSecondaryIndexUpdate>>, /// <p>The new provisioned throughput settings for the specified table or index.</p> #[serde(rename = "ProvisionedThroughput")] #[serde(skip_serializing_if = "Option::is_none")] pub provisioned_throughput: Option<ProvisionedThroughput>, /// <p>The new server-side encryption settings for the specified table.</p> #[serde(rename = "SSESpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub sse_specification: Option<SSESpecification>, /// <p><p>Represents the DynamoDB Streams configuration for the table.</p> <note> <p>You will receive a <code>ResourceInUseException</code> if you attempt to enable a stream on a table that already has a stream, or if you attempt to disable a stream on a table which does not have a stream.</p> </note></p> #[serde(rename = "StreamSpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub stream_specification: Option<StreamSpecification>, /// <p>The name of the table to be updated.</p> #[serde(rename = "TableName")] pub table_name: String, } /// <p>Represents the output of an <code>UpdateTable</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateTableOutput { /// <p>Represents the properties of the table.</p> #[serde(rename = "TableDescription")] #[serde(skip_serializing_if = "Option::is_none")] pub table_description: Option<TableDescription>, } /// <p>Represents the input of an <code>UpdateTimeToLive</code> operation.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize)] pub struct UpdateTimeToLiveInput { /// <p>The name of the table to be configured.</p> #[serde(rename = "TableName")] pub table_name: String, /// <p>Represents the settings used to enable or disable Time to Live for the specified table.</p> #[serde(rename = "TimeToLiveSpecification")] pub time_to_live_specification: TimeToLiveSpecification, } #[derive(Default, Debug, Clone, PartialEq, Deserialize)] #[cfg_attr(test, derive(Serialize))] pub struct UpdateTimeToLiveOutput { /// <p>Represents the output of an <code>UpdateTimeToLive</code> operation.</p> #[serde(rename = "TimeToLiveSpecification")] #[serde(skip_serializing_if = "Option::is_none")] pub time_to_live_specification: Option<TimeToLiveSpecification>, } /// <p>Represents an operation to perform - either <code>DeleteItem</code> or <code>PutItem</code>. You can only request one of these operations, not both, in a single <code>WriteRequest</code>. If you do need to perform both of these operations, you will need to provide two separate <code>WriteRequest</code> objects.</p> #[derive(Default, Debug, Clone, PartialEq, Serialize, Deserialize)] pub struct WriteRequest { /// <p>A request to perform a <code>DeleteItem</code> operation.</p> #[serde(rename = "DeleteRequest")] #[serde(skip_serializing_if = "Option::is_none")] pub delete_request: Option<DeleteRequest>, /// <p>A request to perform a <code>PutItem</code> operation.</p> #[serde(rename = "PutRequest")] #[serde(skip_serializing_if = "Option::is_none")] pub put_request: Option<PutRequest>, } /// Errors returned by BatchGetItem #[derive(Debug, PartialEq)] pub enum BatchGetItemError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl BatchGetItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<BatchGetItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(BatchGetItemError::InternalServerError(err.msg)) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(BatchGetItemError::ProvisionedThroughputExceeded( err.msg, )) } "RequestLimitExceeded" => { return RusotoError::Service(BatchGetItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(BatchGetItemError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for BatchGetItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for BatchGetItemError { fn description(&self) -> &str { match *self { BatchGetItemError::InternalServerError(ref cause) => cause, BatchGetItemError::ProvisionedThroughputExceeded(ref cause) => cause, BatchGetItemError::RequestLimitExceeded(ref cause) => cause, BatchGetItemError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by BatchWriteItem #[derive(Debug, PartialEq)] pub enum BatchWriteItemError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>An item collection is too large. This exception is only returned for tables that have one or more local secondary indexes.</p> ItemCollectionSizeLimitExceeded(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl BatchWriteItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<BatchWriteItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(BatchWriteItemError::InternalServerError(err.msg)) } "ItemCollectionSizeLimitExceededException" => { return RusotoError::Service( BatchWriteItemError::ItemCollectionSizeLimitExceeded(err.msg), ) } "ProvisionedThroughputExceededException" => { return RusotoError::Service( BatchWriteItemError::ProvisionedThroughputExceeded(err.msg), ) } "RequestLimitExceeded" => { return RusotoError::Service(BatchWriteItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(BatchWriteItemError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for BatchWriteItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for BatchWriteItemError { fn description(&self) -> &str { match *self { BatchWriteItemError::InternalServerError(ref cause) => cause, BatchWriteItemError::ItemCollectionSizeLimitExceeded(ref cause) => cause, BatchWriteItemError::ProvisionedThroughputExceeded(ref cause) => cause, BatchWriteItemError::RequestLimitExceeded(ref cause) => cause, BatchWriteItemError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by CreateBackup #[derive(Debug, PartialEq)] pub enum CreateBackupError { /// <p>There is another ongoing conflicting backup control plane operation on the table. The backup is either being created, deleted or restored to a table.</p> BackupInUse(String), /// <p>Backups have not yet been enabled for this table.</p> ContinuousBackupsUnavailable(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>A target table with the specified name is either being created or deleted. </p> TableInUse(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl CreateBackupError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<CreateBackupError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "BackupInUseException" => { return RusotoError::Service(CreateBackupError::BackupInUse(err.msg)) } "ContinuousBackupsUnavailableException" => { return RusotoError::Service(CreateBackupError::ContinuousBackupsUnavailable( err.msg, )) } "InternalServerError" => { return RusotoError::Service(CreateBackupError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(CreateBackupError::LimitExceeded(err.msg)) } "TableInUseException" => { return RusotoError::Service(CreateBackupError::TableInUse(err.msg)) } "TableNotFoundException" => { return RusotoError::Service(CreateBackupError::TableNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for CreateBackupError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for CreateBackupError { fn description(&self) -> &str { match *self { CreateBackupError::BackupInUse(ref cause) => cause, CreateBackupError::ContinuousBackupsUnavailable(ref cause) => cause, CreateBackupError::InternalServerError(ref cause) => cause, CreateBackupError::LimitExceeded(ref cause) => cause, CreateBackupError::TableInUse(ref cause) => cause, CreateBackupError::TableNotFound(ref cause) => cause, } } } /// Errors returned by CreateGlobalTable #[derive(Debug, PartialEq)] pub enum CreateGlobalTableError { /// <p>The specified global table already exists.</p> GlobalTableAlreadyExists(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl CreateGlobalTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<CreateGlobalTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "GlobalTableAlreadyExistsException" => { return RusotoError::Service(CreateGlobalTableError::GlobalTableAlreadyExists( err.msg, )) } "InternalServerError" => { return RusotoError::Service(CreateGlobalTableError::InternalServerError( err.msg, )) } "LimitExceededException" => { return RusotoError::Service(CreateGlobalTableError::LimitExceeded(err.msg)) } "TableNotFoundException" => { return RusotoError::Service(CreateGlobalTableError::TableNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for CreateGlobalTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for CreateGlobalTableError { fn description(&self) -> &str { match *self { CreateGlobalTableError::GlobalTableAlreadyExists(ref cause) => cause, CreateGlobalTableError::InternalServerError(ref cause) => cause, CreateGlobalTableError::LimitExceeded(ref cause) => cause, CreateGlobalTableError::TableNotFound(ref cause) => cause, } } } /// Errors returned by CreateTable #[derive(Debug, PartialEq)] pub enum CreateTableError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), } impl CreateTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<CreateTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(CreateTableError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(CreateTableError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(CreateTableError::ResourceInUse(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for CreateTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for CreateTableError { fn description(&self) -> &str { match *self { CreateTableError::InternalServerError(ref cause) => cause, CreateTableError::LimitExceeded(ref cause) => cause, CreateTableError::ResourceInUse(ref cause) => cause, } } } /// Errors returned by DeleteBackup #[derive(Debug, PartialEq)] pub enum DeleteBackupError { /// <p>There is another ongoing conflicting backup control plane operation on the table. The backup is either being created, deleted or restored to a table.</p> BackupInUse(String), /// <p>Backup not found for the given BackupARN. </p> BackupNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), } impl DeleteBackupError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DeleteBackupError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "BackupInUseException" => { return RusotoError::Service(DeleteBackupError::BackupInUse(err.msg)) } "BackupNotFoundException" => { return RusotoError::Service(DeleteBackupError::BackupNotFound(err.msg)) } "InternalServerError" => { return RusotoError::Service(DeleteBackupError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(DeleteBackupError::LimitExceeded(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DeleteBackupError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DeleteBackupError { fn description(&self) -> &str { match *self { DeleteBackupError::BackupInUse(ref cause) => cause, DeleteBackupError::BackupNotFound(ref cause) => cause, DeleteBackupError::InternalServerError(ref cause) => cause, DeleteBackupError::LimitExceeded(ref cause) => cause, } } } /// Errors returned by DeleteItem #[derive(Debug, PartialEq)] pub enum DeleteItemError { /// <p>A condition specified in the operation could not be evaluated.</p> ConditionalCheckFailed(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>An item collection is too large. This exception is only returned for tables that have one or more local secondary indexes.</p> ItemCollectionSizeLimitExceeded(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), /// <p>Operation was rejected because there is an ongoing transaction for the item.</p> TransactionConflict(String), } impl DeleteItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DeleteItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "ConditionalCheckFailedException" => { return RusotoError::Service(DeleteItemError::ConditionalCheckFailed(err.msg)) } "InternalServerError" => { return RusotoError::Service(DeleteItemError::InternalServerError(err.msg)) } "ItemCollectionSizeLimitExceededException" => { return RusotoError::Service(DeleteItemError::ItemCollectionSizeLimitExceeded( err.msg, )) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(DeleteItemError::ProvisionedThroughputExceeded( err.msg, )) } "RequestLimitExceeded" => { return RusotoError::Service(DeleteItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(DeleteItemError::ResourceNotFound(err.msg)) } "TransactionConflictException" => { return RusotoError::Service(DeleteItemError::TransactionConflict(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DeleteItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DeleteItemError { fn description(&self) -> &str { match *self { DeleteItemError::ConditionalCheckFailed(ref cause) => cause, DeleteItemError::InternalServerError(ref cause) => cause, DeleteItemError::ItemCollectionSizeLimitExceeded(ref cause) => cause, DeleteItemError::ProvisionedThroughputExceeded(ref cause) => cause, DeleteItemError::RequestLimitExceeded(ref cause) => cause, DeleteItemError::ResourceNotFound(ref cause) => cause, DeleteItemError::TransactionConflict(ref cause) => cause, } } } /// Errors returned by DeleteTable #[derive(Debug, PartialEq)] pub enum DeleteTableError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl DeleteTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DeleteTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(DeleteTableError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(DeleteTableError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(DeleteTableError::ResourceInUse(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(DeleteTableError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DeleteTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DeleteTableError { fn description(&self) -> &str { match *self { DeleteTableError::InternalServerError(ref cause) => cause, DeleteTableError::LimitExceeded(ref cause) => cause, DeleteTableError::ResourceInUse(ref cause) => cause, DeleteTableError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by DescribeBackup #[derive(Debug, PartialEq)] pub enum DescribeBackupError { /// <p>Backup not found for the given BackupARN. </p> BackupNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl DescribeBackupError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeBackupError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "BackupNotFoundException" => { return RusotoError::Service(DescribeBackupError::BackupNotFound(err.msg)) } "InternalServerError" => { return RusotoError::Service(DescribeBackupError::InternalServerError(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeBackupError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeBackupError { fn description(&self) -> &str { match *self { DescribeBackupError::BackupNotFound(ref cause) => cause, DescribeBackupError::InternalServerError(ref cause) => cause, } } } /// Errors returned by DescribeContinuousBackups #[derive(Debug, PartialEq)] pub enum DescribeContinuousBackupsError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl DescribeContinuousBackupsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeContinuousBackupsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service( DescribeContinuousBackupsError::InternalServerError(err.msg), ) } "TableNotFoundException" => { return RusotoError::Service(DescribeContinuousBackupsError::TableNotFound( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeContinuousBackupsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeContinuousBackupsError { fn description(&self) -> &str { match *self { DescribeContinuousBackupsError::InternalServerError(ref cause) => cause, DescribeContinuousBackupsError::TableNotFound(ref cause) => cause, } } } /// Errors returned by DescribeEndpoints #[derive(Debug, PartialEq)] pub enum DescribeEndpointsError {} impl DescribeEndpointsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeEndpointsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeEndpointsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeEndpointsError { fn description(&self) -> &str { match *self {} } } /// Errors returned by DescribeGlobalTable #[derive(Debug, PartialEq)] pub enum DescribeGlobalTableError { /// <p>The specified global table does not exist.</p> GlobalTableNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl DescribeGlobalTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeGlobalTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "GlobalTableNotFoundException" => { return RusotoError::Service(DescribeGlobalTableError::GlobalTableNotFound( err.msg, )) } "InternalServerError" => { return RusotoError::Service(DescribeGlobalTableError::InternalServerError( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeGlobalTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeGlobalTableError { fn description(&self) -> &str { match *self { DescribeGlobalTableError::GlobalTableNotFound(ref cause) => cause, DescribeGlobalTableError::InternalServerError(ref cause) => cause, } } } /// Errors returned by DescribeGlobalTableSettings #[derive(Debug, PartialEq)] pub enum DescribeGlobalTableSettingsError { /// <p>The specified global table does not exist.</p> GlobalTableNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl DescribeGlobalTableSettingsError { pub fn from_response( res: BufferedHttpResponse, ) -> RusotoError<DescribeGlobalTableSettingsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "GlobalTableNotFoundException" => { return RusotoError::Service( DescribeGlobalTableSettingsError::GlobalTableNotFound(err.msg), ) } "InternalServerError" => { return RusotoError::Service( DescribeGlobalTableSettingsError::InternalServerError(err.msg), ) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeGlobalTableSettingsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeGlobalTableSettingsError { fn description(&self) -> &str { match *self { DescribeGlobalTableSettingsError::GlobalTableNotFound(ref cause) => cause, DescribeGlobalTableSettingsError::InternalServerError(ref cause) => cause, } } } /// Errors returned by DescribeLimits #[derive(Debug, PartialEq)] pub enum DescribeLimitsError { /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl DescribeLimitsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeLimitsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(DescribeLimitsError::InternalServerError(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeLimitsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeLimitsError { fn description(&self) -> &str { match *self { DescribeLimitsError::InternalServerError(ref cause) => cause, } } } /// Errors returned by DescribeTable #[derive(Debug, PartialEq)] pub enum DescribeTableError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl DescribeTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(DescribeTableError::InternalServerError(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(DescribeTableError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeTableError { fn description(&self) -> &str { match *self { DescribeTableError::InternalServerError(ref cause) => cause, DescribeTableError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by DescribeTimeToLive #[derive(Debug, PartialEq)] pub enum DescribeTimeToLiveError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl DescribeTimeToLiveError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<DescribeTimeToLiveError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(DescribeTimeToLiveError::InternalServerError( err.msg, )) } "ResourceNotFoundException" => { return RusotoError::Service(DescribeTimeToLiveError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for DescribeTimeToLiveError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for DescribeTimeToLiveError { fn description(&self) -> &str { match *self { DescribeTimeToLiveError::InternalServerError(ref cause) => cause, DescribeTimeToLiveError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by GetItem #[derive(Debug, PartialEq)] pub enum GetItemError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl GetItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<GetItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(GetItemError::InternalServerError(err.msg)) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(GetItemError::ProvisionedThroughputExceeded( err.msg, )) } "RequestLimitExceeded" => { return RusotoError::Service(GetItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(GetItemError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for GetItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for GetItemError { fn description(&self) -> &str { match *self { GetItemError::InternalServerError(ref cause) => cause, GetItemError::ProvisionedThroughputExceeded(ref cause) => cause, GetItemError::RequestLimitExceeded(ref cause) => cause, GetItemError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by ListBackups #[derive(Debug, PartialEq)] pub enum ListBackupsError { /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl ListBackupsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<ListBackupsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(ListBackupsError::InternalServerError(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for ListBackupsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for ListBackupsError { fn description(&self) -> &str { match *self { ListBackupsError::InternalServerError(ref cause) => cause, } } } /// Errors returned by ListGlobalTables #[derive(Debug, PartialEq)] pub enum ListGlobalTablesError { /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl ListGlobalTablesError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<ListGlobalTablesError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(ListGlobalTablesError::InternalServerError( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for ListGlobalTablesError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for ListGlobalTablesError { fn description(&self) -> &str { match *self { ListGlobalTablesError::InternalServerError(ref cause) => cause, } } } /// Errors returned by ListTables #[derive(Debug, PartialEq)] pub enum ListTablesError { /// <p>An error occurred on the server side.</p> InternalServerError(String), } impl ListTablesError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<ListTablesError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(ListTablesError::InternalServerError(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for ListTablesError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for ListTablesError { fn description(&self) -> &str { match *self { ListTablesError::InternalServerError(ref cause) => cause, } } } /// Errors returned by ListTagsOfResource #[derive(Debug, PartialEq)] pub enum ListTagsOfResourceError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl ListTagsOfResourceError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<ListTagsOfResourceError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(ListTagsOfResourceError::InternalServerError( err.msg, )) } "ResourceNotFoundException" => { return RusotoError::Service(ListTagsOfResourceError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for ListTagsOfResourceError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for ListTagsOfResourceError { fn description(&self) -> &str { match *self { ListTagsOfResourceError::InternalServerError(ref cause) => cause, ListTagsOfResourceError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by PutItem #[derive(Debug, PartialEq)] pub enum PutItemError { /// <p>A condition specified in the operation could not be evaluated.</p> ConditionalCheckFailed(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>An item collection is too large. This exception is only returned for tables that have one or more local secondary indexes.</p> ItemCollectionSizeLimitExceeded(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), /// <p>Operation was rejected because there is an ongoing transaction for the item.</p> TransactionConflict(String), } impl PutItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<PutItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "ConditionalCheckFailedException" => { return RusotoError::Service(PutItemError::ConditionalCheckFailed(err.msg)) } "InternalServerError" => { return RusotoError::Service(PutItemError::InternalServerError(err.msg)) } "ItemCollectionSizeLimitExceededException" => { return RusotoError::Service(PutItemError::ItemCollectionSizeLimitExceeded( err.msg, )) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(PutItemError::ProvisionedThroughputExceeded( err.msg, )) } "RequestLimitExceeded" => { return RusotoError::Service(PutItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(PutItemError::ResourceNotFound(err.msg)) } "TransactionConflictException" => { return RusotoError::Service(PutItemError::TransactionConflict(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for PutItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for PutItemError { fn description(&self) -> &str { match *self { PutItemError::ConditionalCheckFailed(ref cause) => cause, PutItemError::InternalServerError(ref cause) => cause, PutItemError::ItemCollectionSizeLimitExceeded(ref cause) => cause, PutItemError::ProvisionedThroughputExceeded(ref cause) => cause, PutItemError::RequestLimitExceeded(ref cause) => cause, PutItemError::ResourceNotFound(ref cause) => cause, PutItemError::TransactionConflict(ref cause) => cause, } } } /// Errors returned by Query #[derive(Debug, PartialEq)] pub enum QueryError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl QueryError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<QueryError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(QueryError::InternalServerError(err.msg)) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(QueryError::ProvisionedThroughputExceeded(err.msg)) } "RequestLimitExceeded" => { return RusotoError::Service(QueryError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(QueryError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for QueryError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for QueryError { fn description(&self) -> &str { match *self { QueryError::InternalServerError(ref cause) => cause, QueryError::ProvisionedThroughputExceeded(ref cause) => cause, QueryError::RequestLimitExceeded(ref cause) => cause, QueryError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by RestoreTableFromBackup #[derive(Debug, PartialEq)] pub enum RestoreTableFromBackupError { /// <p>There is another ongoing conflicting backup control plane operation on the table. The backup is either being created, deleted or restored to a table.</p> BackupInUse(String), /// <p>Backup not found for the given BackupARN. </p> BackupNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>A target table with the specified name already exists. </p> TableAlreadyExists(String), /// <p>A target table with the specified name is either being created or deleted. </p> TableInUse(String), } impl RestoreTableFromBackupError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<RestoreTableFromBackupError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "BackupInUseException" => { return RusotoError::Service(RestoreTableFromBackupError::BackupInUse(err.msg)) } "BackupNotFoundException" => { return RusotoError::Service(RestoreTableFromBackupError::BackupNotFound( err.msg, )) } "InternalServerError" => { return RusotoError::Service(RestoreTableFromBackupError::InternalServerError( err.msg, )) } "LimitExceededException" => { return RusotoError::Service(RestoreTableFromBackupError::LimitExceeded( err.msg, )) } "TableAlreadyExistsException" => { return RusotoError::Service(RestoreTableFromBackupError::TableAlreadyExists( err.msg, )) } "TableInUseException" => { return RusotoError::Service(RestoreTableFromBackupError::TableInUse(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for RestoreTableFromBackupError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for RestoreTableFromBackupError { fn description(&self) -> &str { match *self { RestoreTableFromBackupError::BackupInUse(ref cause) => cause, RestoreTableFromBackupError::BackupNotFound(ref cause) => cause, RestoreTableFromBackupError::InternalServerError(ref cause) => cause, RestoreTableFromBackupError::LimitExceeded(ref cause) => cause, RestoreTableFromBackupError::TableAlreadyExists(ref cause) => cause, RestoreTableFromBackupError::TableInUse(ref cause) => cause, } } } /// Errors returned by RestoreTableToPointInTime #[derive(Debug, PartialEq)] pub enum RestoreTableToPointInTimeError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>An invalid restore time was specified. RestoreDateTime must be between EarliestRestorableDateTime and LatestRestorableDateTime.</p> InvalidRestoreTime(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>Point in time recovery has not yet been enabled for this source table.</p> PointInTimeRecoveryUnavailable(String), /// <p>A target table with the specified name already exists. </p> TableAlreadyExists(String), /// <p>A target table with the specified name is either being created or deleted. </p> TableInUse(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl RestoreTableToPointInTimeError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<RestoreTableToPointInTimeError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service( RestoreTableToPointInTimeError::InternalServerError(err.msg), ) } "InvalidRestoreTimeException" => { return RusotoError::Service( RestoreTableToPointInTimeError::InvalidRestoreTime(err.msg), ) } "LimitExceededException" => { return RusotoError::Service(RestoreTableToPointInTimeError::LimitExceeded( err.msg, )) } "PointInTimeRecoveryUnavailableException" => { return RusotoError::Service( RestoreTableToPointInTimeError::PointInTimeRecoveryUnavailable(err.msg), ) } "TableAlreadyExistsException" => { return RusotoError::Service( RestoreTableToPointInTimeError::TableAlreadyExists(err.msg), ) } "TableInUseException" => { return RusotoError::Service(RestoreTableToPointInTimeError::TableInUse( err.msg, )) } "TableNotFoundException" => { return RusotoError::Service(RestoreTableToPointInTimeError::TableNotFound( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for RestoreTableToPointInTimeError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for RestoreTableToPointInTimeError { fn description(&self) -> &str { match *self { RestoreTableToPointInTimeError::InternalServerError(ref cause) => cause, RestoreTableToPointInTimeError::InvalidRestoreTime(ref cause) => cause, RestoreTableToPointInTimeError::LimitExceeded(ref cause) => cause, RestoreTableToPointInTimeError::PointInTimeRecoveryUnavailable(ref cause) => cause, RestoreTableToPointInTimeError::TableAlreadyExists(ref cause) => cause, RestoreTableToPointInTimeError::TableInUse(ref cause) => cause, RestoreTableToPointInTimeError::TableNotFound(ref cause) => cause, } } } /// Errors returned by Scan #[derive(Debug, PartialEq)] pub enum ScanError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl ScanError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<ScanError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(ScanError::InternalServerError(err.msg)) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(ScanError::ProvisionedThroughputExceeded(err.msg)) } "RequestLimitExceeded" => { return RusotoError::Service(ScanError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(ScanError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for ScanError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for ScanError { fn description(&self) -> &str { match *self { ScanError::InternalServerError(ref cause) => cause, ScanError::ProvisionedThroughputExceeded(ref cause) => cause, ScanError::RequestLimitExceeded(ref cause) => cause, ScanError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by TagResource #[derive(Debug, PartialEq)] pub enum TagResourceError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl TagResourceError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<TagResourceError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(TagResourceError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(TagResourceError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(TagResourceError::ResourceInUse(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(TagResourceError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for TagResourceError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for TagResourceError { fn description(&self) -> &str { match *self { TagResourceError::InternalServerError(ref cause) => cause, TagResourceError::LimitExceeded(ref cause) => cause, TagResourceError::ResourceInUse(ref cause) => cause, TagResourceError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by TransactGetItems #[derive(Debug, PartialEq)] pub enum TransactGetItemsError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), /// <p><p>The entire transaction request was canceled.</p> <p>DynamoDB cancels a <code>TransactWriteItems</code> request under the following circumstances:</p> <ul> <li> <p>A condition in one of the condition expressions is not met.</p> </li> <li> <p>A table in the <code>TransactWriteItems</code> request is in a different account or region.</p> </li> <li> <p>More than one action in the <code>TransactWriteItems</code> operation targets the same item.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>An item size becomes too large (larger than 400 KB), or a local secondary index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul> <p>DynamoDB cancels a <code>TransactGetItems</code> request under the following circumstances:</p> <ul> <li> <p>There is an ongoing <code>TransactGetItems</code> operation that conflicts with a concurrent <code>PutItem</code>, <code>UpdateItem</code>, <code>DeleteItem</code> or <code>TransactWriteItems</code> request. In this case the <code>TransactGetItems</code> operation fails with a <code>TransactionCanceledException</code>.</p> </li> <li> <p>A table in the <code>TransactGetItems</code> request is in a different account or region.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul> <note> <p>If using Java, DynamoDB lists the cancellation reasons on the <code>CancellationReasons</code> property. This property is not set for other languages. Transaction cancellation reasons are ordered in the order of requested items, if an item has no error it will have <code>NONE</code> code and <code>Null</code> message.</p> </note> <p>Cancellation reason codes and possible error messages:</p> <ul> <li> <p>No Errors:</p> <ul> <li> <p>Code: <code>NONE</code> </p> </li> <li> <p>Message: <code>null</code> </p> </li> </ul> </li> <li> <p>Conditional Check Failed:</p> <ul> <li> <p>Code: <code>ConditionalCheckFailed</code> </p> </li> <li> <p>Message: The conditional request failed. </p> </li> </ul> </li> <li> <p>Item Collection Size Limit Exceeded:</p> <ul> <li> <p>Code: <code>ItemCollectionSizeLimitExceeded</code> </p> </li> <li> <p>Message: Collection size exceeded.</p> </li> </ul> </li> <li> <p>Transaction Conflict:</p> <ul> <li> <p>Code: <code>TransactionConflict</code> </p> </li> <li> <p>Message: Transaction is ongoing for the item.</p> </li> </ul> </li> <li> <p>Provisioned Throughput Exceeded:</p> <ul> <li> <p>Code: <code>ProvisionedThroughputExceeded</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>The level of configured provisioned throughput for the table was exceeded. Consider increasing your provisioning level with the UpdateTable API.</p> <note> <p>This Message is received when provisioned throughput is exceeded is on a provisioned DynamoDB table.</p> </note> </li> <li> <p>The level of configured provisioned throughput for one or more global secondary indexes of the table was exceeded. Consider increasing your provisioning level for the under-provisioned global secondary indexes with the UpdateTable API.</p> <note> <p>This message is returned when provisioned throughput is exceeded is on a provisioned GSI.</p> </note> </li> </ul> </li> </ul> </li> <li> <p>Throttling Error:</p> <ul> <li> <p>Code: <code>ThrottlingError</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>Throughput exceeds the current capacity of your table or index. DynamoDB is automatically scaling your table or index so please try again shortly. If exceptions persist, check if you have a hot key: https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/bp-partition-key-design.html.</p> <note> <p>This message is returned when writes get throttled on an On-Demand table as DynamoDB is automatically scaling the table.</p> </note> </li> <li> <p>Throughput exceeds the current capacity for one or more global secondary indexes. DynamoDB is automatically scaling your index so please try again shortly.</p> <note> <p>This message is returned when when writes get throttled on an On-Demand GSI as DynamoDB is automatically scaling the GSI.</p> </note> </li> </ul> </li> </ul> </li> <li> <p>Validation Error:</p> <ul> <li> <p>Code: <code>ValidationError</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>One or more parameter values were invalid.</p> </li> <li> <p>The update expression attempted to update the secondary index key beyond allowed size limits.</p> </li> <li> <p>The update expression attempted to update the secondary index key to unsupported type.</p> </li> <li> <p>An operand in the update expression has an incorrect data type.</p> </li> <li> <p>Item size to update has exceeded the maximum allowed size.</p> </li> <li> <p>Number overflow. Attempting to store a number with magnitude larger than supported range.</p> </li> <li> <p>Type mismatch for attribute to update.</p> </li> <li> <p>Nesting Levels have exceeded supported limits.</p> </li> <li> <p>The document path provided in the update expression is invalid for update.</p> </li> <li> <p>The provided expression refers to an attribute that does not exist in the item.</p> </li> </ul> </li> </ul> </li> </ul></p> TransactionCanceled(String), } impl TransactGetItemsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<TransactGetItemsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(TransactGetItemsError::InternalServerError( err.msg, )) } "ProvisionedThroughputExceededException" => { return RusotoError::Service( TransactGetItemsError::ProvisionedThroughputExceeded(err.msg), ) } "ResourceNotFoundException" => { return RusotoError::Service(TransactGetItemsError::ResourceNotFound(err.msg)) } "TransactionCanceledException" => { return RusotoError::Service(TransactGetItemsError::TransactionCanceled( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for TransactGetItemsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for TransactGetItemsError { fn description(&self) -> &str { match *self { TransactGetItemsError::InternalServerError(ref cause) => cause, TransactGetItemsError::ProvisionedThroughputExceeded(ref cause) => cause, TransactGetItemsError::ResourceNotFound(ref cause) => cause, TransactGetItemsError::TransactionCanceled(ref cause) => cause, } } } /// Errors returned by TransactWriteItems #[derive(Debug, PartialEq)] pub enum TransactWriteItemsError { /// <p>DynamoDB rejected the request because you retried a request with a different payload but with an idempotent token that was already used.</p> IdempotentParameterMismatch(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), /// <p><p>The entire transaction request was canceled.</p> <p>DynamoDB cancels a <code>TransactWriteItems</code> request under the following circumstances:</p> <ul> <li> <p>A condition in one of the condition expressions is not met.</p> </li> <li> <p>A table in the <code>TransactWriteItems</code> request is in a different account or region.</p> </li> <li> <p>More than one action in the <code>TransactWriteItems</code> operation targets the same item.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>An item size becomes too large (larger than 400 KB), or a local secondary index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul> <p>DynamoDB cancels a <code>TransactGetItems</code> request under the following circumstances:</p> <ul> <li> <p>There is an ongoing <code>TransactGetItems</code> operation that conflicts with a concurrent <code>PutItem</code>, <code>UpdateItem</code>, <code>DeleteItem</code> or <code>TransactWriteItems</code> request. In this case the <code>TransactGetItems</code> operation fails with a <code>TransactionCanceledException</code>.</p> </li> <li> <p>A table in the <code>TransactGetItems</code> request is in a different account or region.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul> <note> <p>If using Java, DynamoDB lists the cancellation reasons on the <code>CancellationReasons</code> property. This property is not set for other languages. Transaction cancellation reasons are ordered in the order of requested items, if an item has no error it will have <code>NONE</code> code and <code>Null</code> message.</p> </note> <p>Cancellation reason codes and possible error messages:</p> <ul> <li> <p>No Errors:</p> <ul> <li> <p>Code: <code>NONE</code> </p> </li> <li> <p>Message: <code>null</code> </p> </li> </ul> </li> <li> <p>Conditional Check Failed:</p> <ul> <li> <p>Code: <code>ConditionalCheckFailed</code> </p> </li> <li> <p>Message: The conditional request failed. </p> </li> </ul> </li> <li> <p>Item Collection Size Limit Exceeded:</p> <ul> <li> <p>Code: <code>ItemCollectionSizeLimitExceeded</code> </p> </li> <li> <p>Message: Collection size exceeded.</p> </li> </ul> </li> <li> <p>Transaction Conflict:</p> <ul> <li> <p>Code: <code>TransactionConflict</code> </p> </li> <li> <p>Message: Transaction is ongoing for the item.</p> </li> </ul> </li> <li> <p>Provisioned Throughput Exceeded:</p> <ul> <li> <p>Code: <code>ProvisionedThroughputExceeded</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>The level of configured provisioned throughput for the table was exceeded. Consider increasing your provisioning level with the UpdateTable API.</p> <note> <p>This Message is received when provisioned throughput is exceeded is on a provisioned DynamoDB table.</p> </note> </li> <li> <p>The level of configured provisioned throughput for one or more global secondary indexes of the table was exceeded. Consider increasing your provisioning level for the under-provisioned global secondary indexes with the UpdateTable API.</p> <note> <p>This message is returned when provisioned throughput is exceeded is on a provisioned GSI.</p> </note> </li> </ul> </li> </ul> </li> <li> <p>Throttling Error:</p> <ul> <li> <p>Code: <code>ThrottlingError</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>Throughput exceeds the current capacity of your table or index. DynamoDB is automatically scaling your table or index so please try again shortly. If exceptions persist, check if you have a hot key: https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/bp-partition-key-design.html.</p> <note> <p>This message is returned when writes get throttled on an On-Demand table as DynamoDB is automatically scaling the table.</p> </note> </li> <li> <p>Throughput exceeds the current capacity for one or more global secondary indexes. DynamoDB is automatically scaling your index so please try again shortly.</p> <note> <p>This message is returned when when writes get throttled on an On-Demand GSI as DynamoDB is automatically scaling the GSI.</p> </note> </li> </ul> </li> </ul> </li> <li> <p>Validation Error:</p> <ul> <li> <p>Code: <code>ValidationError</code> </p> </li> <li> <p>Messages: </p> <ul> <li> <p>One or more parameter values were invalid.</p> </li> <li> <p>The update expression attempted to update the secondary index key beyond allowed size limits.</p> </li> <li> <p>The update expression attempted to update the secondary index key to unsupported type.</p> </li> <li> <p>An operand in the update expression has an incorrect data type.</p> </li> <li> <p>Item size to update has exceeded the maximum allowed size.</p> </li> <li> <p>Number overflow. Attempting to store a number with magnitude larger than supported range.</p> </li> <li> <p>Type mismatch for attribute to update.</p> </li> <li> <p>Nesting Levels have exceeded supported limits.</p> </li> <li> <p>The document path provided in the update expression is invalid for update.</p> </li> <li> <p>The provided expression refers to an attribute that does not exist in the item.</p> </li> </ul> </li> </ul> </li> </ul></p> TransactionCanceled(String), /// <p>The transaction with the given request token is already in progress.</p> TransactionInProgress(String), } impl TransactWriteItemsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<TransactWriteItemsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "IdempotentParameterMismatchException" => { return RusotoError::Service( TransactWriteItemsError::IdempotentParameterMismatch(err.msg), ) } "InternalServerError" => { return RusotoError::Service(TransactWriteItemsError::InternalServerError( err.msg, )) } "ProvisionedThroughputExceededException" => { return RusotoError::Service( TransactWriteItemsError::ProvisionedThroughputExceeded(err.msg), ) } "ResourceNotFoundException" => { return RusotoError::Service(TransactWriteItemsError::ResourceNotFound(err.msg)) } "TransactionCanceledException" => { return RusotoError::Service(TransactWriteItemsError::TransactionCanceled( err.msg, )) } "TransactionInProgressException" => { return RusotoError::Service(TransactWriteItemsError::TransactionInProgress( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for TransactWriteItemsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for TransactWriteItemsError { fn description(&self) -> &str { match *self { TransactWriteItemsError::IdempotentParameterMismatch(ref cause) => cause, TransactWriteItemsError::InternalServerError(ref cause) => cause, TransactWriteItemsError::ProvisionedThroughputExceeded(ref cause) => cause, TransactWriteItemsError::ResourceNotFound(ref cause) => cause, TransactWriteItemsError::TransactionCanceled(ref cause) => cause, TransactWriteItemsError::TransactionInProgress(ref cause) => cause, } } } /// Errors returned by UntagResource #[derive(Debug, PartialEq)] pub enum UntagResourceError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl UntagResourceError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UntagResourceError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(UntagResourceError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(UntagResourceError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(UntagResourceError::ResourceInUse(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(UntagResourceError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UntagResourceError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UntagResourceError { fn description(&self) -> &str { match *self { UntagResourceError::InternalServerError(ref cause) => cause, UntagResourceError::LimitExceeded(ref cause) => cause, UntagResourceError::ResourceInUse(ref cause) => cause, UntagResourceError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by UpdateContinuousBackups #[derive(Debug, PartialEq)] pub enum UpdateContinuousBackupsError { /// <p>Backups have not yet been enabled for this table.</p> ContinuousBackupsUnavailable(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl UpdateContinuousBackupsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateContinuousBackupsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "ContinuousBackupsUnavailableException" => { return RusotoError::Service( UpdateContinuousBackupsError::ContinuousBackupsUnavailable(err.msg), ) } "InternalServerError" => { return RusotoError::Service(UpdateContinuousBackupsError::InternalServerError( err.msg, )) } "TableNotFoundException" => { return RusotoError::Service(UpdateContinuousBackupsError::TableNotFound( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateContinuousBackupsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateContinuousBackupsError { fn description(&self) -> &str { match *self { UpdateContinuousBackupsError::ContinuousBackupsUnavailable(ref cause) => cause, UpdateContinuousBackupsError::InternalServerError(ref cause) => cause, UpdateContinuousBackupsError::TableNotFound(ref cause) => cause, } } } /// Errors returned by UpdateGlobalTable #[derive(Debug, PartialEq)] pub enum UpdateGlobalTableError { /// <p>The specified global table does not exist.</p> GlobalTableNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>The specified replica is already part of the global table.</p> ReplicaAlreadyExists(String), /// <p>The specified replica is no longer part of the global table.</p> ReplicaNotFound(String), /// <p>A source table with the name <code>TableName</code> does not currently exist within the subscriber's account.</p> TableNotFound(String), } impl UpdateGlobalTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateGlobalTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "GlobalTableNotFoundException" => { return RusotoError::Service(UpdateGlobalTableError::GlobalTableNotFound( err.msg, )) } "InternalServerError" => { return RusotoError::Service(UpdateGlobalTableError::InternalServerError( err.msg, )) } "ReplicaAlreadyExistsException" => { return RusotoError::Service(UpdateGlobalTableError::ReplicaAlreadyExists( err.msg, )) } "ReplicaNotFoundException" => { return RusotoError::Service(UpdateGlobalTableError::ReplicaNotFound(err.msg)) } "TableNotFoundException" => { return RusotoError::Service(UpdateGlobalTableError::TableNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateGlobalTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateGlobalTableError { fn description(&self) -> &str { match *self { UpdateGlobalTableError::GlobalTableNotFound(ref cause) => cause, UpdateGlobalTableError::InternalServerError(ref cause) => cause, UpdateGlobalTableError::ReplicaAlreadyExists(ref cause) => cause, UpdateGlobalTableError::ReplicaNotFound(ref cause) => cause, UpdateGlobalTableError::TableNotFound(ref cause) => cause, } } } /// Errors returned by UpdateGlobalTableSettings #[derive(Debug, PartialEq)] pub enum UpdateGlobalTableSettingsError { /// <p>The specified global table does not exist.</p> GlobalTableNotFound(String), /// <p>The operation tried to access a nonexistent index.</p> IndexNotFound(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The specified replica is no longer part of the global table.</p> ReplicaNotFound(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), } impl UpdateGlobalTableSettingsError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateGlobalTableSettingsError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "GlobalTableNotFoundException" => { return RusotoError::Service( UpdateGlobalTableSettingsError::GlobalTableNotFound(err.msg), ) } "IndexNotFoundException" => { return RusotoError::Service(UpdateGlobalTableSettingsError::IndexNotFound( err.msg, )) } "InternalServerError" => { return RusotoError::Service( UpdateGlobalTableSettingsError::InternalServerError(err.msg), ) } "LimitExceededException" => { return RusotoError::Service(UpdateGlobalTableSettingsError::LimitExceeded( err.msg, )) } "ReplicaNotFoundException" => { return RusotoError::Service(UpdateGlobalTableSettingsError::ReplicaNotFound( err.msg, )) } "ResourceInUseException" => { return RusotoError::Service(UpdateGlobalTableSettingsError::ResourceInUse( err.msg, )) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateGlobalTableSettingsError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateGlobalTableSettingsError { fn description(&self) -> &str { match *self { UpdateGlobalTableSettingsError::GlobalTableNotFound(ref cause) => cause, UpdateGlobalTableSettingsError::IndexNotFound(ref cause) => cause, UpdateGlobalTableSettingsError::InternalServerError(ref cause) => cause, UpdateGlobalTableSettingsError::LimitExceeded(ref cause) => cause, UpdateGlobalTableSettingsError::ReplicaNotFound(ref cause) => cause, UpdateGlobalTableSettingsError::ResourceInUse(ref cause) => cause, } } } /// Errors returned by UpdateItem #[derive(Debug, PartialEq)] pub enum UpdateItemError { /// <p>A condition specified in the operation could not be evaluated.</p> ConditionalCheckFailed(String), /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>An item collection is too large. This exception is only returned for tables that have one or more local secondary indexes.</p> ItemCollectionSizeLimitExceeded(String), /// <p>Your request rate is too high. The AWS SDKs for DynamoDB automatically retry requests that receive this exception. Your request is eventually successful, unless your retry queue is too large to finish. Reduce the frequency of requests and use exponential backoff. For more information, go to <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Programming.Errors.html#Programming.Errors.RetryAndBackoff">Error Retries and Exponential Backoff</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> ProvisionedThroughputExceeded(String), /// <p>Throughput exceeds the current throughput limit for your account. Please contact AWS Support at <a href="https://aws.amazon.com/support">AWS Support</a> to request a limit increase.</p> RequestLimitExceeded(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), /// <p>Operation was rejected because there is an ongoing transaction for the item.</p> TransactionConflict(String), } impl UpdateItemError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateItemError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "ConditionalCheckFailedException" => { return RusotoError::Service(UpdateItemError::ConditionalCheckFailed(err.msg)) } "InternalServerError" => { return RusotoError::Service(UpdateItemError::InternalServerError(err.msg)) } "ItemCollectionSizeLimitExceededException" => { return RusotoError::Service(UpdateItemError::ItemCollectionSizeLimitExceeded( err.msg, )) } "ProvisionedThroughputExceededException" => { return RusotoError::Service(UpdateItemError::ProvisionedThroughputExceeded( err.msg, )) } "RequestLimitExceeded" => { return RusotoError::Service(UpdateItemError::RequestLimitExceeded(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(UpdateItemError::ResourceNotFound(err.msg)) } "TransactionConflictException" => { return RusotoError::Service(UpdateItemError::TransactionConflict(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateItemError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateItemError { fn description(&self) -> &str { match *self { UpdateItemError::ConditionalCheckFailed(ref cause) => cause, UpdateItemError::InternalServerError(ref cause) => cause, UpdateItemError::ItemCollectionSizeLimitExceeded(ref cause) => cause, UpdateItemError::ProvisionedThroughputExceeded(ref cause) => cause, UpdateItemError::RequestLimitExceeded(ref cause) => cause, UpdateItemError::ResourceNotFound(ref cause) => cause, UpdateItemError::TransactionConflict(ref cause) => cause, } } } /// Errors returned by UpdateTable #[derive(Debug, PartialEq)] pub enum UpdateTableError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl UpdateTableError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateTableError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(UpdateTableError::InternalServerError(err.msg)) } "LimitExceededException" => { return RusotoError::Service(UpdateTableError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(UpdateTableError::ResourceInUse(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(UpdateTableError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateTableError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateTableError { fn description(&self) -> &str { match *self { UpdateTableError::InternalServerError(ref cause) => cause, UpdateTableError::LimitExceeded(ref cause) => cause, UpdateTableError::ResourceInUse(ref cause) => cause, UpdateTableError::ResourceNotFound(ref cause) => cause, } } } /// Errors returned by UpdateTimeToLive #[derive(Debug, PartialEq)] pub enum UpdateTimeToLiveError { /// <p>An error occurred on the server side.</p> InternalServerError(String), /// <p>There is no limit to the number of daily on-demand backups that can be taken. </p> <p>Up to 50 simultaneous table operations are allowed per account. These operations include <code>CreateTable</code>, <code>UpdateTable</code>, <code>DeleteTable</code>,<code>UpdateTimeToLive</code>, <code>RestoreTableFromBackup</code>, and <code>RestoreTableToPointInTime</code>. </p> <p>The only exception is when you are creating a table with one or more secondary indexes. You can have up to 25 such requests running at a time; however, if the table or index specifications are complex, DynamoDB might temporarily reduce the number of concurrent operations.</p> <p>There is a soft account limit of 256 tables.</p> LimitExceeded(String), /// <p>The operation conflicts with the resource's availability. For example, you attempted to recreate an existing table, or tried to delete a table currently in the <code>CREATING</code> state.</p> ResourceInUse(String), /// <p>The operation tried to access a nonexistent table or index. The resource might not be specified correctly, or its status might not be <code>ACTIVE</code>.</p> ResourceNotFound(String), } impl UpdateTimeToLiveError { pub fn from_response(res: BufferedHttpResponse) -> RusotoError<UpdateTimeToLiveError> { if let Some(err) = proto::json::Error::parse(&res) { match err.typ.as_str() { "InternalServerError" => { return RusotoError::Service(UpdateTimeToLiveError::InternalServerError( err.msg, )) } "LimitExceededException" => { return RusotoError::Service(UpdateTimeToLiveError::LimitExceeded(err.msg)) } "ResourceInUseException" => { return RusotoError::Service(UpdateTimeToLiveError::ResourceInUse(err.msg)) } "ResourceNotFoundException" => { return RusotoError::Service(UpdateTimeToLiveError::ResourceNotFound(err.msg)) } "ValidationException" => return RusotoError::Validation(err.msg), _ => {} } } return RusotoError::Unknown(res); } } impl fmt::Display for UpdateTimeToLiveError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{}", self.description()) } } impl Error for UpdateTimeToLiveError { fn description(&self) -> &str { match *self { UpdateTimeToLiveError::InternalServerError(ref cause) => cause, UpdateTimeToLiveError::LimitExceeded(ref cause) => cause, UpdateTimeToLiveError::ResourceInUse(ref cause) => cause, UpdateTimeToLiveError::ResourceNotFound(ref cause) => cause, } } } /// Trait representing the capabilities of the DynamoDB API. DynamoDB clients implement this trait. pub trait DynamoDb { /// <p>The <code>BatchGetItem</code> operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key.</p> <p>A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. <code>BatchGetItem</code> will return a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for <code>UnprocessedKeys</code>. You can use this value to retry the operation starting with the next item to get.</p> <important> <p>If you request more than 100 items <code>BatchGetItem</code> will return a <code>ValidationException</code> with the message "Too many items requested for the BatchGetItem call".</p> </important> <p>For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate <code>UnprocessedKeys</code> value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one data set.</p> <p>If <i>none</i> of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then <code>BatchGetItem</code> will return a <code>ProvisionedThroughputExceededException</code>. If <i>at least one</i> of the items is successfully processed, then <code>BatchGetItem</code> completes successfully, while returning the keys of the unread items in <code>UnprocessedKeys</code>.</p> <important> <p>If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, <i>we strongly recommend that you use an exponential backoff algorithm</i>. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ErrorHandling.html#BatchOperations">Batch Operations and Error Handling</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </important> <p>By default, <code>BatchGetItem</code> performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set <code>ConsistentRead</code> to <code>true</code> for any or all tables.</p> <p>In order to minimize response latency, <code>BatchGetItem</code> retrieves items in parallel.</p> <p>When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the <code>ProjectionExpression</code> parameter.</p> <p>If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#CapacityUnitCalculations">Capacity Units Calculations</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn batch_get_item( &self, input: BatchGetItemInput, ) -> RusotoFuture<BatchGetItemOutput, BatchGetItemError>; /// <p><p>The <code>BatchWriteItem</code> operation puts or deletes multiple items in one or more tables. A single call to <code>BatchWriteItem</code> can write up to 16 MB of data, which can comprise as many as 25 put or delete requests. Individual items to be written can be as large as 400 KB.</p> <note> <p> <code>BatchWriteItem</code> cannot update items. To update items, use the <code>UpdateItem</code> action.</p> </note> <p>The individual <code>PutItem</code> and <code>DeleteItem</code> operations specified in <code>BatchWriteItem</code> are atomic; however <code>BatchWriteItem</code> as a whole is not. If any requested operations fail because the table's provisioned throughput is exceeded or an internal processing failure occurs, the failed operations are returned in the <code>UnprocessedItems</code> response parameter. You can investigate and optionally resend the requests. Typically, you would call <code>BatchWriteItem</code> in a loop. Each iteration would check for unprocessed items and submit a new <code>BatchWriteItem</code> request with those unprocessed items until all items have been processed.</p> <p>Note that if <i>none</i> of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then <code>BatchWriteItem</code> will return a <code>ProvisionedThroughputExceededException</code>.</p> <important> <p>If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, <i>we strongly recommend that you use an exponential backoff algorithm</i>. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ErrorHandling.html#BatchOperations">Batch Operations and Error Handling</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </important> <p>With <code>BatchWriteItem</code>, you can efficiently write or delete large amounts of data, such as from Amazon Elastic MapReduce (EMR), or copy data from another database into DynamoDB. In order to improve performance with these large-scale operations, <code>BatchWriteItem</code> does not behave in the same way as individual <code>PutItem</code> and <code>DeleteItem</code> calls would. For example, you cannot specify conditions on individual put and delete requests, and <code>BatchWriteItem</code> does not return deleted items in the response.</p> <p>If you use a programming language that supports concurrency, you can use threads to write items in parallel. Your application must include the necessary logic to manage the threads. With languages that don't support threading, you must update or delete the specified items one at a time. In both situations, <code>BatchWriteItem</code> performs the specified put and delete operations in parallel, giving you the power of the thread pool approach without having to introduce complexity into your application.</p> <p>Parallel processing reduces latency, but each specified put and delete request consumes the same number of write capacity units whether it is processed in parallel or not. Delete operations on nonexistent items consume one write capacity unit.</p> <p>If one or more of the following is true, DynamoDB rejects the entire batch write operation:</p> <ul> <li> <p>One or more tables specified in the <code>BatchWriteItem</code> request does not exist.</p> </li> <li> <p>Primary key attributes specified on an item in the request do not match those in the corresponding table's primary key schema.</p> </li> <li> <p>You try to perform multiple operations on the same item in the same <code>BatchWriteItem</code> request. For example, you cannot put and delete the same item in the same <code>BatchWriteItem</code> request. </p> </li> <li> <p> Your request contains at least two items with identical hash and range keys (which essentially is two put operations). </p> </li> <li> <p>There are more than 25 requests in the batch.</p> </li> <li> <p>Any individual item in a batch exceeds 400 KB.</p> </li> <li> <p>The total request size exceeds 16 MB.</p> </li> </ul></p> fn batch_write_item( &self, input: BatchWriteItemInput, ) -> RusotoFuture<BatchWriteItemOutput, BatchWriteItemError>; /// <p><p>Creates a backup for an existing table.</p> <p> Each time you create an On-Demand Backup, the entire table data is backed up. There is no limit to the number of on-demand backups that can be taken. </p> <p> When you create an On-Demand Backup, a time marker of the request is cataloged, and the backup is created asynchronously, by applying all changes until the time of the request to the last full table snapshot. Backup requests are processed instantaneously and become available for restore within minutes. </p> <p>You can call <code>CreateBackup</code> at a maximum rate of 50 times per second.</p> <p>All backups in DynamoDB work without consuming any provisioned throughput on the table.</p> <p> If you submit a backup request on 2018-12-14 at 14:25:00, the backup is guaranteed to contain all data committed to the table up to 14:24:00, and data committed after 14:26:00 will not be. The backup may or may not contain data modifications made between 14:24:00 and 14:26:00. On-Demand Backup does not support causal consistency. </p> <p> Along with data, the following are also included on the backups: </p> <ul> <li> <p>Global secondary indexes (GSIs)</p> </li> <li> <p>Local secondary indexes (LSIs)</p> </li> <li> <p>Streams</p> </li> <li> <p>Provisioned read and write capacity</p> </li> </ul></p> fn create_backup( &self, input: CreateBackupInput, ) -> RusotoFuture<CreateBackupOutput, CreateBackupError>; /// <p><p>Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided regions. </p> <p>If you want to add a new replica table to a global table, each of the following conditions must be true:</p> <ul> <li> <p>The table must have the same primary key as all of the other replicas.</p> </li> <li> <p>The table must have the same name as all of the other replicas.</p> </li> <li> <p>The table must have DynamoDB Streams enabled, with the stream containing both the new and the old images of the item.</p> </li> <li> <p>None of the replica tables in the global table can contain any data.</p> </li> </ul> <p> If global secondary indexes are specified, then the following conditions must also be met: </p> <ul> <li> <p> The global secondary indexes must have the same name. </p> </li> <li> <p> The global secondary indexes must have the same hash key and sort key (if present). </p> </li> </ul> <important> <p> Write capacity settings should be set consistently across your replica tables and secondary indexes. DynamoDB strongly recommends enabling auto scaling to manage the write capacity settings for all of your global tables replicas and indexes. </p> <p> If you prefer to manage write capacity settings manually, you should provision equal replicated write capacity units to your replica tables. You should also provision equal replicated write capacity units to matching secondary indexes across your global table. </p> </important></p> fn create_global_table( &self, input: CreateGlobalTableInput, ) -> RusotoFuture<CreateGlobalTableOutput, CreateGlobalTableError>; /// <p>The <code>CreateTable</code> operation adds a new table to your account. In an AWS account, table names must be unique within each region. That is, you can have two tables with same name if you create the tables in different regions.</p> <p> <code>CreateTable</code> is an asynchronous operation. Upon receiving a <code>CreateTable</code> request, DynamoDB immediately returns a response with a <code>TableStatus</code> of <code>CREATING</code>. After the table is created, DynamoDB sets the <code>TableStatus</code> to <code>ACTIVE</code>. You can perform read and write operations only on an <code>ACTIVE</code> table. </p> <p>You can optionally define secondary indexes on the new table, as part of the <code>CreateTable</code> operation. If you want to create multiple tables with secondary indexes on them, you must create the tables sequentially. Only one table with secondary indexes can be in the <code>CREATING</code> state at any given time.</p> <p>You can use the <code>DescribeTable</code> action to check the table status.</p> fn create_table( &self, input: CreateTableInput, ) -> RusotoFuture<CreateTableOutput, CreateTableError>; /// <p>Deletes an existing backup of a table.</p> <p>You can call <code>DeleteBackup</code> at a maximum rate of 10 times per second.</p> fn delete_backup( &self, input: DeleteBackupInput, ) -> RusotoFuture<DeleteBackupOutput, DeleteBackupError>; /// <p>Deletes a single item in a table by primary key. You can perform a conditional delete operation that deletes the item if it exists, or if it has an expected attribute value.</p> <p>In addition to deleting an item, you can also return the item's attribute values in the same operation, using the <code>ReturnValues</code> parameter.</p> <p>Unless you specify conditions, the <code>DeleteItem</code> is an idempotent operation; running it multiple times on the same item or attribute does <i>not</i> result in an error response.</p> <p>Conditional deletes are useful for deleting items only if specific conditions are met. If those conditions are met, DynamoDB performs the delete. Otherwise, the item is not deleted.</p> fn delete_item( &self, input: DeleteItemInput, ) -> RusotoFuture<DeleteItemOutput, DeleteItemError>; /// <p>The <code>DeleteTable</code> operation deletes a table and all of its items. After a <code>DeleteTable</code> request, the specified table is in the <code>DELETING</code> state until DynamoDB completes the deletion. If the table is in the <code>ACTIVE</code> state, you can delete it. If a table is in <code>CREATING</code> or <code>UPDATING</code> states, then DynamoDB returns a <code>ResourceInUseException</code>. If the specified table does not exist, DynamoDB returns a <code>ResourceNotFoundException</code>. If table is already in the <code>DELETING</code> state, no error is returned. </p> <note> <p>DynamoDB might continue to accept data read and write operations, such as <code>GetItem</code> and <code>PutItem</code>, on a table in the <code>DELETING</code> state until the table deletion is complete.</p> </note> <p>When you delete a table, any indexes on that table are also deleted.</p> <p>If you have DynamoDB Streams enabled on the table, then the corresponding stream on that table goes into the <code>DISABLED</code> state, and the stream is automatically deleted after 24 hours.</p> <p>Use the <code>DescribeTable</code> action to check the status of the table. </p> fn delete_table( &self, input: DeleteTableInput, ) -> RusotoFuture<DeleteTableOutput, DeleteTableError>; /// <p>Describes an existing backup of a table.</p> <p>You can call <code>DescribeBackup</code> at a maximum rate of 10 times per second.</p> fn describe_backup( &self, input: DescribeBackupInput, ) -> RusotoFuture<DescribeBackupOutput, DescribeBackupError>; /// <p>Checks the status of continuous backups and point in time recovery on the specified table. Continuous backups are <code>ENABLED</code> on all tables at table creation. If point in time recovery is enabled, <code>PointInTimeRecoveryStatus</code> will be set to ENABLED.</p> <p> Once continuous backups and point in time recovery are enabled, you can restore to any point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. </p> <p> <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days. </p> <p>You can call <code>DescribeContinuousBackups</code> at a maximum rate of 10 times per second.</p> fn describe_continuous_backups( &self, input: DescribeContinuousBackupsInput, ) -> RusotoFuture<DescribeContinuousBackupsOutput, DescribeContinuousBackupsError>; /// <p>Returns the regional endpoint information.</p> fn describe_endpoints(&self) -> RusotoFuture<DescribeEndpointsResponse, DescribeEndpointsError>; /// <p>Returns information about the specified global table.</p> fn describe_global_table( &self, input: DescribeGlobalTableInput, ) -> RusotoFuture<DescribeGlobalTableOutput, DescribeGlobalTableError>; /// <p>Describes region specific settings for a global table.</p> fn describe_global_table_settings( &self, input: DescribeGlobalTableSettingsInput, ) -> RusotoFuture<DescribeGlobalTableSettingsOutput, DescribeGlobalTableSettingsError>; /// <p>Returns the current provisioned-capacity limits for your AWS account in a region, both for the region as a whole and for any one DynamoDB table that you create there.</p> <p>When you establish an AWS account, the account has initial limits on the maximum read capacity units and write capacity units that you can provision across all of your DynamoDB tables in a given region. Also, there are per-table limits that apply when you create a table there. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> page in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>Although you can increase these limits by filing a case at <a href="https://console.aws.amazon.com/support/home#/">AWS Support Center</a>, obtaining the increase is not instantaneous. The <code>DescribeLimits</code> action lets you write code to compare the capacity you are currently using to those limits imposed by your account so that you have enough time to apply for an increase before you hit a limit.</p> <p>For example, you could use one of the AWS SDKs to do the following:</p> <ol> <li> <p>Call <code>DescribeLimits</code> for a particular region to obtain your current account limits on provisioned capacity there.</p> </li> <li> <p>Create a variable to hold the aggregate read capacity units provisioned for all your tables in that region, and one to hold the aggregate write capacity units. Zero them both.</p> </li> <li> <p>Call <code>ListTables</code> to obtain a list of all your DynamoDB tables.</p> </li> <li> <p>For each table name listed by <code>ListTables</code>, do the following:</p> <ul> <li> <p>Call <code>DescribeTable</code> with the table name.</p> </li> <li> <p>Use the data returned by <code>DescribeTable</code> to add the read capacity units and write capacity units provisioned for the table itself to your variables.</p> </li> <li> <p>If the table has one or more global secondary indexes (GSIs), loop over these GSIs and add their provisioned capacity values to your variables as well.</p> </li> </ul> </li> <li> <p>Report the account limits for that region returned by <code>DescribeLimits</code>, along with the total current provisioned capacity levels you have calculated.</p> </li> </ol> <p>This will let you see whether you are getting close to your account-level limits.</p> <p>The per-table limits apply only when you are creating a new table. They restrict the sum of the provisioned capacity of the new table itself and all its global secondary indexes.</p> <p>For existing tables and their GSIs, DynamoDB will not let you increase provisioned capacity extremely rapidly, but the only upper limit that applies is that the aggregate provisioned capacity over all your tables and GSIs cannot exceed either of the per-account limits.</p> <note> <p> <code>DescribeLimits</code> should only be called periodically. You can expect throttling errors if you call it more than once in a minute.</p> </note> <p>The <code>DescribeLimits</code> Request element has no content.</p> fn describe_limits(&self) -> RusotoFuture<DescribeLimitsOutput, DescribeLimitsError>; /// <p><p>Returns information about the table, including the current status of the table, when it was created, the primary key schema, and any indexes on the table.</p> <note> <p>If you issue a <code>DescribeTable</code> request immediately after a <code>CreateTable</code> request, DynamoDB might return a <code>ResourceNotFoundException</code>. This is because <code>DescribeTable</code> uses an eventually consistent query, and the metadata for your table might not be available at that moment. Wait for a few seconds, and then try the <code>DescribeTable</code> request again.</p> </note></p> fn describe_table( &self, input: DescribeTableInput, ) -> RusotoFuture<DescribeTableOutput, DescribeTableError>; /// <p>Gives a description of the Time to Live (TTL) status on the specified table. </p> fn describe_time_to_live( &self, input: DescribeTimeToLiveInput, ) -> RusotoFuture<DescribeTimeToLiveOutput, DescribeTimeToLiveError>; /// <p>The <code>GetItem</code> operation returns a set of attributes for the item with the given primary key. If there is no matching item, <code>GetItem</code> does not return any data and there will be no <code>Item</code> element in the response.</p> <p> <code>GetItem</code> provides an eventually consistent read by default. If your application requires a strongly consistent read, set <code>ConsistentRead</code> to <code>true</code>. Although a strongly consistent read might take more time than an eventually consistent read, it always returns the last updated value.</p> fn get_item(&self, input: GetItemInput) -> RusotoFuture<GetItemOutput, GetItemError>; /// <p>List backups associated with an AWS account. To list backups for a given table, specify <code>TableName</code>. <code>ListBackups</code> returns a paginated list of results with at most 1MB worth of items in a page. You can also specify a limit for the maximum number of entries to be returned in a page. </p> <p>In the request, start time is inclusive but end time is exclusive. Note that these limits are for the time at which the original backup was requested.</p> <p>You can call <code>ListBackups</code> a maximum of 5 times per second.</p> fn list_backups( &self, input: ListBackupsInput, ) -> RusotoFuture<ListBackupsOutput, ListBackupsError>; /// <p>Lists all global tables that have a replica in the specified region.</p> fn list_global_tables( &self, input: ListGlobalTablesInput, ) -> RusotoFuture<ListGlobalTablesOutput, ListGlobalTablesError>; /// <p>Returns an array of table names associated with the current account and endpoint. The output from <code>ListTables</code> is paginated, with each page returning a maximum of 100 table names.</p> fn list_tables( &self, input: ListTablesInput, ) -> RusotoFuture<ListTablesOutput, ListTablesError>; /// <p>List all tags on an Amazon DynamoDB resource. You can call ListTagsOfResource up to 10 times per second, per account.</p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn list_tags_of_resource( &self, input: ListTagsOfResourceInput, ) -> RusotoFuture<ListTagsOfResourceOutput, ListTagsOfResourceError>; /// <p>Creates a new item, or replaces an old item with a new item. If an item that has the same primary key as the new item already exists in the specified table, the new item completely replaces the existing item. You can perform a conditional put operation (add a new item if one with the specified primary key doesn't exist), or replace an existing item if it has certain attribute values. You can return the item's attribute values in the same operation, using the <code>ReturnValues</code> parameter.</p> <important> <p>This topic provides general information about the <code>PutItem</code> API.</p> <p>For information on how to call the <code>PutItem</code> API using the AWS SDK in specific languages, see the following:</p> <ul> <li> <p> <a href="http://docs.aws.amazon.com/goto/aws-cli/dynamodb-2012-08-10/PutItem"> PutItem in the AWS Command Line Interface </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/DotNetSDKV3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for .NET </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForCpp/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for C++ </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForGoV1/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Go </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForJava/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Java </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/AWSJavaScriptSDK/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for JavaScript </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForPHPV3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for PHP V3 </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/boto3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Python </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForRubyV2/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Ruby V2 </a> </p> </li> </ul> </important> <p>When you add an item, the primary key attribute(s) are the only required attributes. Attribute values cannot be null. String and Binary type attributes must have lengths greater than zero. Set type attributes cannot be empty. Requests with empty values will be rejected with a <code>ValidationException</code> exception.</p> <note> <p>To prevent a new item from replacing an existing item, use a conditional expression that contains the <code>attribute_not_exists</code> function with the name of the attribute being used as the partition key for the table. Since every record must contain that attribute, the <code>attribute_not_exists</code> function will only succeed if no matching item exists.</p> </note> <p>For more information about <code>PutItem</code>, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithItems.html">Working with Items</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn put_item(&self, input: PutItemInput) -> RusotoFuture<PutItemOutput, PutItemError>; /// <p>The <code>Query</code> operation finds items based on primary key values. You can query any table or secondary index that has a composite primary key (a partition key and a sort key). </p> <p>Use the <code>KeyConditionExpression</code> parameter to provide a specific value for the partition key. The <code>Query</code> operation will return all of the items from the table or index with that partition key value. You can optionally narrow the scope of the <code>Query</code> operation by specifying a sort key value and a comparison operator in <code>KeyConditionExpression</code>. To further refine the <code>Query</code> results, you can optionally provide a <code>FilterExpression</code>. A <code>FilterExpression</code> determines which items within the results should be returned to you. All of the other results are discarded. </p> <p> A <code>Query</code> operation always returns a result set. If no matching items are found, the result set will be empty. Queries that do not return results consume the minimum number of read capacity units for that type of read operation. </p> <note> <p> DynamoDB calculates the number of read capacity units consumed based on item size, not on the amount of data that is returned to an application. The number of capacity units consumed will be the same whether you request all of the attributes (the default behavior) or just some of them (using a projection expression). The number will also be the same whether or not you use a <code>FilterExpression</code>. </p> </note> <p> <code>Query</code> results are always sorted by the sort key value. If the data type of the sort key is Number, the results are returned in numeric order; otherwise, the results are returned in order of UTF-8 bytes. By default, the sort order is ascending. To reverse the order, set the <code>ScanIndexForward</code> parameter to false. </p> <p> A single <code>Query</code> operation will read up to the maximum number of items set (if using the <code>Limit</code> parameter) or a maximum of 1 MB of data and then apply any filtering to the results using <code>FilterExpression</code>. If <code>LastEvaluatedKey</code> is present in the response, you will need to paginate the result set. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Query.html#Query.Pagination">Paginating the Results</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> <p> <code>FilterExpression</code> is applied after a <code>Query</code> finishes, but before the results are returned. A <code>FilterExpression</code> cannot contain partition key or sort key attributes. You need to specify those attributes in the <code>KeyConditionExpression</code>. </p> <note> <p> A <code>Query</code> operation can return an empty result set and a <code>LastEvaluatedKey</code> if all the items read for the page of results are filtered out. </p> </note> <p>You can query a table, a local secondary index, or a global secondary index. For a query on a table or on a local secondary index, you can set the <code>ConsistentRead</code> parameter to <code>true</code> and obtain a strongly consistent result. Global secondary indexes support eventually consistent reads only, so do not specify <code>ConsistentRead</code> when querying a global secondary index.</p> fn query(&self, input: QueryInput) -> RusotoFuture<QueryOutput, QueryError>; /// <p><p>Creates a new table from an existing backup. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. </p> <p>You can call <code>RestoreTableFromBackup</code> at a maximum rate of 10 times per second.</p> <p>You must manually set up the following on the restored table:</p> <ul> <li> <p>Auto scaling policies</p> </li> <li> <p>IAM policies</p> </li> <li> <p>Cloudwatch metrics and alarms</p> </li> <li> <p>Tags</p> </li> <li> <p>Stream settings</p> </li> <li> <p>Time to Live (TTL) settings</p> </li> </ul></p> fn restore_table_from_backup( &self, input: RestoreTableFromBackupInput, ) -> RusotoFuture<RestoreTableFromBackupOutput, RestoreTableFromBackupError>; /// <p><p>Restores the specified table to the specified point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. You can restore your table to any point in time during the last 35 days. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. </p> <p> When you restore using point in time recovery, DynamoDB restores your table data to the state based on the selected date and time (day:hour:minute:second) to a new table. </p> <p> Along with data, the following are also included on the new restored table using point in time recovery: </p> <ul> <li> <p>Global secondary indexes (GSIs)</p> </li> <li> <p>Local secondary indexes (LSIs)</p> </li> <li> <p>Provisioned read and write capacity</p> </li> <li> <p>Encryption settings</p> <important> <p> All these settings come from the current settings of the source table at the time of restore. </p> </important> </li> </ul> <p>You must manually set up the following on the restored table:</p> <ul> <li> <p>Auto scaling policies</p> </li> <li> <p>IAM policies</p> </li> <li> <p>Cloudwatch metrics and alarms</p> </li> <li> <p>Tags</p> </li> <li> <p>Stream settings</p> </li> <li> <p>Time to Live (TTL) settings</p> </li> <li> <p>Point in time recovery settings</p> </li> </ul></p> fn restore_table_to_point_in_time( &self, input: RestoreTableToPointInTimeInput, ) -> RusotoFuture<RestoreTableToPointInTimeOutput, RestoreTableToPointInTimeError>; /// <p>The <code>Scan</code> operation returns one or more items and item attributes by accessing every item in a table or a secondary index. To have DynamoDB return fewer items, you can provide a <code>FilterExpression</code> operation.</p> <p>If the total number of scanned items exceeds the maximum data set size limit of 1 MB, the scan stops and results are returned to the user as a <code>LastEvaluatedKey</code> value to continue the scan in a subsequent operation. The results also include the number of items exceeding the limit. A scan can result in no table data meeting the filter criteria. </p> <p>A single <code>Scan</code> operation will read up to the maximum number of items set (if using the <code>Limit</code> parameter) or a maximum of 1 MB of data and then apply any filtering to the results using <code>FilterExpression</code>. If <code>LastEvaluatedKey</code> is present in the response, you will need to paginate the result set. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Scan.html#Scan.Pagination">Paginating the Results</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> <p> <code>Scan</code> operations proceed sequentially; however, for faster performance on a large table or secondary index, applications can request a parallel <code>Scan</code> operation by providing the <code>Segment</code> and <code>TotalSegments</code> parameters. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Scan.html#Scan.ParallelScan">Parallel Scan</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p> <code>Scan</code> uses eventually consistent reads when accessing the data in a table; therefore, the result set might not include the changes to data in the table immediately before the operation began. If you need a consistent copy of the data, as of the time that the <code>Scan</code> begins, you can set the <code>ConsistentRead</code> parameter to <code>true</code>.</p> fn scan(&self, input: ScanInput) -> RusotoFuture<ScanOutput, ScanError>; /// <p>Associate a set of tags with an Amazon DynamoDB resource. You can then activate these user-defined tags so that they appear on the Billing and Cost Management console for cost allocation tracking. You can call TagResource up to 5 times per second, per account. </p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn tag_resource(&self, input: TagResourceInput) -> RusotoFuture<(), TagResourceError>; /// <p><p> <code>TransactGetItems</code> is a synchronous operation that atomically retrieves multiple items from one or more tables (but not from indexes) in a single account and region. A <code>TransactGetItems</code> call can contain up to 10 <code>TransactGetItem</code> objects, each of which contains a <code>Get</code> structure that specifies an item to retrieve from a table in the account and region. A call to <code>TransactGetItems</code> cannot retrieve items from tables in more than one AWS account or region.</p> <p>DynamoDB rejects the entire <code>TransactGetItems</code> request if any of the following is true:</p> <ul> <li> <p>A conflicting operation is in the process of updating an item to be read.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul></p> fn transact_get_items( &self, input: TransactGetItemsInput, ) -> RusotoFuture<TransactGetItemsOutput, TransactGetItemsError>; /// <p><p> <code>TransactWriteItems</code> is a synchronous write operation that groups up to 10 action requests. These actions can target items in different tables, but not in different AWS accounts or regions, and no two actions can target the same item. For example, you cannot both <code>ConditionCheck</code> and <code>Update</code> the same item.</p> <p>The actions are completed atomically so that either all of them succeed, or all of them fail. They are defined by the following objects:</p> <ul> <li> <p> <code>Put</code> — Initiates a <code>PutItem</code> operation to write a new item. This structure specifies the primary key of the item to be written, the name of the table to write it in, an optional condition expression that must be satisfied for the write to succeed, a list of the item's attributes, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>Update</code> — Initiates an <code>UpdateItem</code> operation to update an existing item. This structure specifies the primary key of the item to be updated, the name of the table where it resides, an optional condition expression that must be satisfied for the update to succeed, an expression that defines one or more attributes to be updated, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>Delete</code> — Initiates a <code>DeleteItem</code> operation to delete an existing item. This structure specifies the primary key of the item to be deleted, the name of the table where it resides, an optional condition expression that must be satisfied for the deletion to succeed, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>ConditionCheck</code> — Applies a condition to an item that is not being modified by the transaction. This structure specifies the primary key of the item to be checked, the name of the table where it resides, a condition expression that must be satisfied for the transaction to succeed, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> </ul> <p>DynamoDB rejects the entire <code>TransactWriteItems</code> request if any of the following is true:</p> <ul> <li> <p>A condition in one of the condition expressions is not met.</p> </li> <li> <p>A conflicting operation is in the process of updating the same item.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>An item size becomes too large (bigger than 400 KB), a Local Secondary Index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul></p> fn transact_write_items( &self, input: TransactWriteItemsInput, ) -> RusotoFuture<TransactWriteItemsOutput, TransactWriteItemsError>; /// <p>Removes the association of tags from an Amazon DynamoDB resource. You can call UntagResource up to 5 times per second, per account. </p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn untag_resource(&self, input: UntagResourceInput) -> RusotoFuture<(), UntagResourceError>; /// <p> <code>UpdateContinuousBackups</code> enables or disables point in time recovery for the specified table. A successful <code>UpdateContinuousBackups</code> call returns the current <code>ContinuousBackupsDescription</code>. Continuous backups are <code>ENABLED</code> on all tables at table creation. If point in time recovery is enabled, <code>PointInTimeRecoveryStatus</code> will be set to ENABLED.</p> <p> Once continuous backups and point in time recovery are enabled, you can restore to any point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. </p> <p> <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days.. </p> fn update_continuous_backups( &self, input: UpdateContinuousBackupsInput, ) -> RusotoFuture<UpdateContinuousBackupsOutput, UpdateContinuousBackupsError>; /// <p><p>Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, must have the same name as the global table, must have the same key schema, and must have DynamoDB Streams enabled and must have same provisioned and maximum write capacity units.</p> <note> <p>Although you can use <code>UpdateGlobalTable</code> to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas.</p> </note> <p> If global secondary indexes are specified, then the following conditions must also be met: </p> <ul> <li> <p> The global secondary indexes must have the same name. </p> </li> <li> <p> The global secondary indexes must have the same hash key and sort key (if present). </p> </li> <li> <p> The global secondary indexes must have the same provisioned and maximum write capacity units. </p> </li> </ul></p> fn update_global_table( &self, input: UpdateGlobalTableInput, ) -> RusotoFuture<UpdateGlobalTableOutput, UpdateGlobalTableError>; /// <p>Updates settings for a global table.</p> fn update_global_table_settings( &self, input: UpdateGlobalTableSettingsInput, ) -> RusotoFuture<UpdateGlobalTableSettingsOutput, UpdateGlobalTableSettingsError>; /// <p>Edits an existing item's attributes, or adds a new item to the table if it does not already exist. You can put, delete, or add attribute values. You can also perform a conditional update on an existing item (insert a new attribute name-value pair if it doesn't exist, or replace an existing name-value pair if it has certain expected attribute values).</p> <p>You can also return the item's attribute values in the same <code>UpdateItem</code> operation using the <code>ReturnValues</code> parameter.</p> fn update_item( &self, input: UpdateItemInput, ) -> RusotoFuture<UpdateItemOutput, UpdateItemError>; /// <p>Modifies the provisioned throughput settings, global secondary indexes, or DynamoDB Streams settings for a given table.</p> <p>You can only perform one of the following operations at once:</p> <ul> <li> <p>Modify the provisioned throughput settings of the table.</p> </li> <li> <p>Enable or disable Streams on the table.</p> </li> <li> <p>Remove a global secondary index from the table.</p> </li> <li> <p>Create a new global secondary index on the table. Once the index begins backfilling, you can use <code>UpdateTable</code> to perform other operations.</p> </li> </ul> <p> <code>UpdateTable</code> is an asynchronous operation; while it is executing, the table status changes from <code>ACTIVE</code> to <code>UPDATING</code>. While it is <code>UPDATING</code>, you cannot issue another <code>UpdateTable</code> request. When the table returns to the <code>ACTIVE</code> state, the <code>UpdateTable</code> operation is complete.</p> fn update_table( &self, input: UpdateTableInput, ) -> RusotoFuture<UpdateTableOutput, UpdateTableError>; /// <p>The UpdateTimeToLive method will enable or disable TTL for the specified table. A successful <code>UpdateTimeToLive</code> call returns the current <code>TimeToLiveSpecification</code>; it may take up to one hour for the change to fully process. Any additional <code>UpdateTimeToLive</code> calls for the same table during this one hour duration result in a <code>ValidationException</code>. </p> <p>TTL compares the current time in epoch time format to the time stored in the TTL attribute of an item. If the epoch time value stored in the attribute is less than the current time, the item is marked as expired and subsequently deleted.</p> <note> <p> The epoch time format is the number of seconds elapsed since 12:00:00 AM January 1st, 1970 UTC. </p> </note> <p>DynamoDB deletes expired items on a best-effort basis to ensure availability of throughput for other data operations. </p> <important> <p>DynamoDB typically deletes expired items within two days of expiration. The exact duration within which an item gets deleted after expiration is specific to the nature of the workload. Items that have expired and not been deleted will still show up in reads, queries, and scans.</p> </important> <p>As items are deleted, they are removed from any Local Secondary Index and Global Secondary Index immediately in the same eventually consistent way as a standard delete operation.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/TTL.html">Time To Live</a> in the Amazon DynamoDB Developer Guide. </p> fn update_time_to_live( &self, input: UpdateTimeToLiveInput, ) -> RusotoFuture<UpdateTimeToLiveOutput, UpdateTimeToLiveError>; } /// A client for the DynamoDB API. #[derive(Clone)] pub struct DynamoDbClient { client: Client, region: region::Region, } impl DynamoDbClient { /// Creates a client backed by the default tokio event loop. /// /// The client will use the default credentials provider and tls client. pub fn new(region: region::Region) -> DynamoDbClient { DynamoDbClient { client: Client::shared(), region, } } pub fn new_with<P, D>( request_dispatcher: D, credentials_provider: P, region: region::Region, ) -> DynamoDbClient where P: ProvideAwsCredentials + Send + Sync + 'static, P::Future: Send, D: DispatchSignedRequest + Send + Sync + 'static, D::Future: Send, { DynamoDbClient { client: Client::new_with(credentials_provider, request_dispatcher), region, } } } impl DynamoDb for DynamoDbClient { /// <p>The <code>BatchGetItem</code> operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key.</p> <p>A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. <code>BatchGetItem</code> will return a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for <code>UnprocessedKeys</code>. You can use this value to retry the operation starting with the next item to get.</p> <important> <p>If you request more than 100 items <code>BatchGetItem</code> will return a <code>ValidationException</code> with the message "Too many items requested for the BatchGetItem call".</p> </important> <p>For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate <code>UnprocessedKeys</code> value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one data set.</p> <p>If <i>none</i> of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then <code>BatchGetItem</code> will return a <code>ProvisionedThroughputExceededException</code>. If <i>at least one</i> of the items is successfully processed, then <code>BatchGetItem</code> completes successfully, while returning the keys of the unread items in <code>UnprocessedKeys</code>.</p> <important> <p>If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, <i>we strongly recommend that you use an exponential backoff algorithm</i>. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ErrorHandling.html#BatchOperations">Batch Operations and Error Handling</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </important> <p>By default, <code>BatchGetItem</code> performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set <code>ConsistentRead</code> to <code>true</code> for any or all tables.</p> <p>In order to minimize response latency, <code>BatchGetItem</code> retrieves items in parallel.</p> <p>When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the <code>ProjectionExpression</code> parameter.</p> <p>If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithTables.html#CapacityUnitCalculations">Capacity Units Calculations</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn batch_get_item( &self, input: BatchGetItemInput, ) -> RusotoFuture<BatchGetItemOutput, BatchGetItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.BatchGetItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<BatchGetItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(BatchGetItemError::from_response(response))), ) } }) } /// <p><p>The <code>BatchWriteItem</code> operation puts or deletes multiple items in one or more tables. A single call to <code>BatchWriteItem</code> can write up to 16 MB of data, which can comprise as many as 25 put or delete requests. Individual items to be written can be as large as 400 KB.</p> <note> <p> <code>BatchWriteItem</code> cannot update items. To update items, use the <code>UpdateItem</code> action.</p> </note> <p>The individual <code>PutItem</code> and <code>DeleteItem</code> operations specified in <code>BatchWriteItem</code> are atomic; however <code>BatchWriteItem</code> as a whole is not. If any requested operations fail because the table's provisioned throughput is exceeded or an internal processing failure occurs, the failed operations are returned in the <code>UnprocessedItems</code> response parameter. You can investigate and optionally resend the requests. Typically, you would call <code>BatchWriteItem</code> in a loop. Each iteration would check for unprocessed items and submit a new <code>BatchWriteItem</code> request with those unprocessed items until all items have been processed.</p> <p>Note that if <i>none</i> of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then <code>BatchWriteItem</code> will return a <code>ProvisionedThroughputExceededException</code>.</p> <important> <p>If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, <i>we strongly recommend that you use an exponential backoff algorithm</i>. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ErrorHandling.html#BatchOperations">Batch Operations and Error Handling</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> </important> <p>With <code>BatchWriteItem</code>, you can efficiently write or delete large amounts of data, such as from Amazon Elastic MapReduce (EMR), or copy data from another database into DynamoDB. In order to improve performance with these large-scale operations, <code>BatchWriteItem</code> does not behave in the same way as individual <code>PutItem</code> and <code>DeleteItem</code> calls would. For example, you cannot specify conditions on individual put and delete requests, and <code>BatchWriteItem</code> does not return deleted items in the response.</p> <p>If you use a programming language that supports concurrency, you can use threads to write items in parallel. Your application must include the necessary logic to manage the threads. With languages that don't support threading, you must update or delete the specified items one at a time. In both situations, <code>BatchWriteItem</code> performs the specified put and delete operations in parallel, giving you the power of the thread pool approach without having to introduce complexity into your application.</p> <p>Parallel processing reduces latency, but each specified put and delete request consumes the same number of write capacity units whether it is processed in parallel or not. Delete operations on nonexistent items consume one write capacity unit.</p> <p>If one or more of the following is true, DynamoDB rejects the entire batch write operation:</p> <ul> <li> <p>One or more tables specified in the <code>BatchWriteItem</code> request does not exist.</p> </li> <li> <p>Primary key attributes specified on an item in the request do not match those in the corresponding table's primary key schema.</p> </li> <li> <p>You try to perform multiple operations on the same item in the same <code>BatchWriteItem</code> request. For example, you cannot put and delete the same item in the same <code>BatchWriteItem</code> request. </p> </li> <li> <p> Your request contains at least two items with identical hash and range keys (which essentially is two put operations). </p> </li> <li> <p>There are more than 25 requests in the batch.</p> </li> <li> <p>Any individual item in a batch exceeds 400 KB.</p> </li> <li> <p>The total request size exceeds 16 MB.</p> </li> </ul></p> fn batch_write_item( &self, input: BatchWriteItemInput, ) -> RusotoFuture<BatchWriteItemOutput, BatchWriteItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.BatchWriteItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<BatchWriteItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(BatchWriteItemError::from_response(response))), ) } }) } /// <p><p>Creates a backup for an existing table.</p> <p> Each time you create an On-Demand Backup, the entire table data is backed up. There is no limit to the number of on-demand backups that can be taken. </p> <p> When you create an On-Demand Backup, a time marker of the request is cataloged, and the backup is created asynchronously, by applying all changes until the time of the request to the last full table snapshot. Backup requests are processed instantaneously and become available for restore within minutes. </p> <p>You can call <code>CreateBackup</code> at a maximum rate of 50 times per second.</p> <p>All backups in DynamoDB work without consuming any provisioned throughput on the table.</p> <p> If you submit a backup request on 2018-12-14 at 14:25:00, the backup is guaranteed to contain all data committed to the table up to 14:24:00, and data committed after 14:26:00 will not be. The backup may or may not contain data modifications made between 14:24:00 and 14:26:00. On-Demand Backup does not support causal consistency. </p> <p> Along with data, the following are also included on the backups: </p> <ul> <li> <p>Global secondary indexes (GSIs)</p> </li> <li> <p>Local secondary indexes (LSIs)</p> </li> <li> <p>Streams</p> </li> <li> <p>Provisioned read and write capacity</p> </li> </ul></p> fn create_backup( &self, input: CreateBackupInput, ) -> RusotoFuture<CreateBackupOutput, CreateBackupError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.CreateBackup"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<CreateBackupOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(CreateBackupError::from_response(response))), ) } }) } /// <p><p>Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided regions. </p> <p>If you want to add a new replica table to a global table, each of the following conditions must be true:</p> <ul> <li> <p>The table must have the same primary key as all of the other replicas.</p> </li> <li> <p>The table must have the same name as all of the other replicas.</p> </li> <li> <p>The table must have DynamoDB Streams enabled, with the stream containing both the new and the old images of the item.</p> </li> <li> <p>None of the replica tables in the global table can contain any data.</p> </li> </ul> <p> If global secondary indexes are specified, then the following conditions must also be met: </p> <ul> <li> <p> The global secondary indexes must have the same name. </p> </li> <li> <p> The global secondary indexes must have the same hash key and sort key (if present). </p> </li> </ul> <important> <p> Write capacity settings should be set consistently across your replica tables and secondary indexes. DynamoDB strongly recommends enabling auto scaling to manage the write capacity settings for all of your global tables replicas and indexes. </p> <p> If you prefer to manage write capacity settings manually, you should provision equal replicated write capacity units to your replica tables. You should also provision equal replicated write capacity units to matching secondary indexes across your global table. </p> </important></p> fn create_global_table( &self, input: CreateGlobalTableInput, ) -> RusotoFuture<CreateGlobalTableOutput, CreateGlobalTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.CreateGlobalTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<CreateGlobalTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(CreateGlobalTableError::from_response(response))), ) } }) } /// <p>The <code>CreateTable</code> operation adds a new table to your account. In an AWS account, table names must be unique within each region. That is, you can have two tables with same name if you create the tables in different regions.</p> <p> <code>CreateTable</code> is an asynchronous operation. Upon receiving a <code>CreateTable</code> request, DynamoDB immediately returns a response with a <code>TableStatus</code> of <code>CREATING</code>. After the table is created, DynamoDB sets the <code>TableStatus</code> to <code>ACTIVE</code>. You can perform read and write operations only on an <code>ACTIVE</code> table. </p> <p>You can optionally define secondary indexes on the new table, as part of the <code>CreateTable</code> operation. If you want to create multiple tables with secondary indexes on them, you must create the tables sequentially. Only one table with secondary indexes can be in the <code>CREATING</code> state at any given time.</p> <p>You can use the <code>DescribeTable</code> action to check the table status.</p> fn create_table( &self, input: CreateTableInput, ) -> RusotoFuture<CreateTableOutput, CreateTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.CreateTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<CreateTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(CreateTableError::from_response(response))), ) } }) } /// <p>Deletes an existing backup of a table.</p> <p>You can call <code>DeleteBackup</code> at a maximum rate of 10 times per second.</p> fn delete_backup( &self, input: DeleteBackupInput, ) -> RusotoFuture<DeleteBackupOutput, DeleteBackupError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DeleteBackup"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DeleteBackupOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DeleteBackupError::from_response(response))), ) } }) } /// <p>Deletes a single item in a table by primary key. You can perform a conditional delete operation that deletes the item if it exists, or if it has an expected attribute value.</p> <p>In addition to deleting an item, you can also return the item's attribute values in the same operation, using the <code>ReturnValues</code> parameter.</p> <p>Unless you specify conditions, the <code>DeleteItem</code> is an idempotent operation; running it multiple times on the same item or attribute does <i>not</i> result in an error response.</p> <p>Conditional deletes are useful for deleting items only if specific conditions are met. If those conditions are met, DynamoDB performs the delete. Otherwise, the item is not deleted.</p> fn delete_item( &self, input: DeleteItemInput, ) -> RusotoFuture<DeleteItemOutput, DeleteItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DeleteItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DeleteItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DeleteItemError::from_response(response))), ) } }) } /// <p>The <code>DeleteTable</code> operation deletes a table and all of its items. After a <code>DeleteTable</code> request, the specified table is in the <code>DELETING</code> state until DynamoDB completes the deletion. If the table is in the <code>ACTIVE</code> state, you can delete it. If a table is in <code>CREATING</code> or <code>UPDATING</code> states, then DynamoDB returns a <code>ResourceInUseException</code>. If the specified table does not exist, DynamoDB returns a <code>ResourceNotFoundException</code>. If table is already in the <code>DELETING</code> state, no error is returned. </p> <note> <p>DynamoDB might continue to accept data read and write operations, such as <code>GetItem</code> and <code>PutItem</code>, on a table in the <code>DELETING</code> state until the table deletion is complete.</p> </note> <p>When you delete a table, any indexes on that table are also deleted.</p> <p>If you have DynamoDB Streams enabled on the table, then the corresponding stream on that table goes into the <code>DISABLED</code> state, and the stream is automatically deleted after 24 hours.</p> <p>Use the <code>DescribeTable</code> action to check the status of the table. </p> fn delete_table( &self, input: DeleteTableInput, ) -> RusotoFuture<DeleteTableOutput, DeleteTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DeleteTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DeleteTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DeleteTableError::from_response(response))), ) } }) } /// <p>Describes an existing backup of a table.</p> <p>You can call <code>DescribeBackup</code> at a maximum rate of 10 times per second.</p> fn describe_backup( &self, input: DescribeBackupInput, ) -> RusotoFuture<DescribeBackupOutput, DescribeBackupError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeBackup"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeBackupOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DescribeBackupError::from_response(response))), ) } }) } /// <p>Checks the status of continuous backups and point in time recovery on the specified table. Continuous backups are <code>ENABLED</code> on all tables at table creation. If point in time recovery is enabled, <code>PointInTimeRecoveryStatus</code> will be set to ENABLED.</p> <p> Once continuous backups and point in time recovery are enabled, you can restore to any point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. </p> <p> <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days. </p> <p>You can call <code>DescribeContinuousBackups</code> at a maximum rate of 10 times per second.</p> fn describe_continuous_backups( &self, input: DescribeContinuousBackupsInput, ) -> RusotoFuture<DescribeContinuousBackupsOutput, DescribeContinuousBackupsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header( "x-amz-target", "DynamoDB_20120810.DescribeContinuousBackups", ); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeContinuousBackupsOutput, _>() })) } else { Box::new(response.buffer().from_err().and_then(|response| { Err(DescribeContinuousBackupsError::from_response(response)) })) } }) } /// <p>Returns the regional endpoint information.</p> fn describe_endpoints( &self, ) -> RusotoFuture<DescribeEndpointsResponse, DescribeEndpointsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeEndpoints"); request.set_payload(Some(bytes::Bytes::from_static(b"{}"))); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeEndpointsResponse, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DescribeEndpointsError::from_response(response))), ) } }) } /// <p>Returns information about the specified global table.</p> fn describe_global_table( &self, input: DescribeGlobalTableInput, ) -> RusotoFuture<DescribeGlobalTableOutput, DescribeGlobalTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeGlobalTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeGlobalTableOutput, _>() })) } else { Box::new( response.buffer().from_err().and_then(|response| { Err(DescribeGlobalTableError::from_response(response)) }), ) } }) } /// <p>Describes region specific settings for a global table.</p> fn describe_global_table_settings( &self, input: DescribeGlobalTableSettingsInput, ) -> RusotoFuture<DescribeGlobalTableSettingsOutput, DescribeGlobalTableSettingsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header( "x-amz-target", "DynamoDB_20120810.DescribeGlobalTableSettings", ); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeGlobalTableSettingsOutput, _>() })) } else { Box::new(response.buffer().from_err().and_then(|response| { Err(DescribeGlobalTableSettingsError::from_response(response)) })) } }) } /// <p>Returns the current provisioned-capacity limits for your AWS account in a region, both for the region as a whole and for any one DynamoDB table that you create there.</p> <p>When you establish an AWS account, the account has initial limits on the maximum read capacity units and write capacity units that you can provision across all of your DynamoDB tables in a given region. Also, there are per-table limits that apply when you create a table there. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html">Limits</a> page in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p>Although you can increase these limits by filing a case at <a href="https://console.aws.amazon.com/support/home#/">AWS Support Center</a>, obtaining the increase is not instantaneous. The <code>DescribeLimits</code> action lets you write code to compare the capacity you are currently using to those limits imposed by your account so that you have enough time to apply for an increase before you hit a limit.</p> <p>For example, you could use one of the AWS SDKs to do the following:</p> <ol> <li> <p>Call <code>DescribeLimits</code> for a particular region to obtain your current account limits on provisioned capacity there.</p> </li> <li> <p>Create a variable to hold the aggregate read capacity units provisioned for all your tables in that region, and one to hold the aggregate write capacity units. Zero them both.</p> </li> <li> <p>Call <code>ListTables</code> to obtain a list of all your DynamoDB tables.</p> </li> <li> <p>For each table name listed by <code>ListTables</code>, do the following:</p> <ul> <li> <p>Call <code>DescribeTable</code> with the table name.</p> </li> <li> <p>Use the data returned by <code>DescribeTable</code> to add the read capacity units and write capacity units provisioned for the table itself to your variables.</p> </li> <li> <p>If the table has one or more global secondary indexes (GSIs), loop over these GSIs and add their provisioned capacity values to your variables as well.</p> </li> </ul> </li> <li> <p>Report the account limits for that region returned by <code>DescribeLimits</code>, along with the total current provisioned capacity levels you have calculated.</p> </li> </ol> <p>This will let you see whether you are getting close to your account-level limits.</p> <p>The per-table limits apply only when you are creating a new table. They restrict the sum of the provisioned capacity of the new table itself and all its global secondary indexes.</p> <p>For existing tables and their GSIs, DynamoDB will not let you increase provisioned capacity extremely rapidly, but the only upper limit that applies is that the aggregate provisioned capacity over all your tables and GSIs cannot exceed either of the per-account limits.</p> <note> <p> <code>DescribeLimits</code> should only be called periodically. You can expect throttling errors if you call it more than once in a minute.</p> </note> <p>The <code>DescribeLimits</code> Request element has no content.</p> fn describe_limits(&self) -> RusotoFuture<DescribeLimitsOutput, DescribeLimitsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeLimits"); request.set_payload(Some(bytes::Bytes::from_static(b"{}"))); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeLimitsOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DescribeLimitsError::from_response(response))), ) } }) } /// <p><p>Returns information about the table, including the current status of the table, when it was created, the primary key schema, and any indexes on the table.</p> <note> <p>If you issue a <code>DescribeTable</code> request immediately after a <code>CreateTable</code> request, DynamoDB might return a <code>ResourceNotFoundException</code>. This is because <code>DescribeTable</code> uses an eventually consistent query, and the metadata for your table might not be available at that moment. Wait for a few seconds, and then try the <code>DescribeTable</code> request again.</p> </note></p> fn describe_table( &self, input: DescribeTableInput, ) -> RusotoFuture<DescribeTableOutput, DescribeTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DescribeTableError::from_response(response))), ) } }) } /// <p>Gives a description of the Time to Live (TTL) status on the specified table. </p> fn describe_time_to_live( &self, input: DescribeTimeToLiveInput, ) -> RusotoFuture<DescribeTimeToLiveOutput, DescribeTimeToLiveError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.DescribeTimeToLive"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<DescribeTimeToLiveOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(DescribeTimeToLiveError::from_response(response))), ) } }) } /// <p>The <code>GetItem</code> operation returns a set of attributes for the item with the given primary key. If there is no matching item, <code>GetItem</code> does not return any data and there will be no <code>Item</code> element in the response.</p> <p> <code>GetItem</code> provides an eventually consistent read by default. If your application requires a strongly consistent read, set <code>ConsistentRead</code> to <code>true</code>. Although a strongly consistent read might take more time than an eventually consistent read, it always returns the last updated value.</p> fn get_item(&self, input: GetItemInput) -> RusotoFuture<GetItemOutput, GetItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.GetItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response).deserialize::<GetItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(GetItemError::from_response(response))), ) } }) } /// <p>List backups associated with an AWS account. To list backups for a given table, specify <code>TableName</code>. <code>ListBackups</code> returns a paginated list of results with at most 1MB worth of items in a page. You can also specify a limit for the maximum number of entries to be returned in a page. </p> <p>In the request, start time is inclusive but end time is exclusive. Note that these limits are for the time at which the original backup was requested.</p> <p>You can call <code>ListBackups</code> a maximum of 5 times per second.</p> fn list_backups( &self, input: ListBackupsInput, ) -> RusotoFuture<ListBackupsOutput, ListBackupsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.ListBackups"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<ListBackupsOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(ListBackupsError::from_response(response))), ) } }) } /// <p>Lists all global tables that have a replica in the specified region.</p> fn list_global_tables( &self, input: ListGlobalTablesInput, ) -> RusotoFuture<ListGlobalTablesOutput, ListGlobalTablesError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.ListGlobalTables"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<ListGlobalTablesOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(ListGlobalTablesError::from_response(response))), ) } }) } /// <p>Returns an array of table names associated with the current account and endpoint. The output from <code>ListTables</code> is paginated, with each page returning a maximum of 100 table names.</p> fn list_tables( &self, input: ListTablesInput, ) -> RusotoFuture<ListTablesOutput, ListTablesError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.ListTables"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<ListTablesOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(ListTablesError::from_response(response))), ) } }) } /// <p>List all tags on an Amazon DynamoDB resource. You can call ListTagsOfResource up to 10 times per second, per account.</p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn list_tags_of_resource( &self, input: ListTagsOfResourceInput, ) -> RusotoFuture<ListTagsOfResourceOutput, ListTagsOfResourceError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.ListTagsOfResource"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<ListTagsOfResourceOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(ListTagsOfResourceError::from_response(response))), ) } }) } /// <p>Creates a new item, or replaces an old item with a new item. If an item that has the same primary key as the new item already exists in the specified table, the new item completely replaces the existing item. You can perform a conditional put operation (add a new item if one with the specified primary key doesn't exist), or replace an existing item if it has certain attribute values. You can return the item's attribute values in the same operation, using the <code>ReturnValues</code> parameter.</p> <important> <p>This topic provides general information about the <code>PutItem</code> API.</p> <p>For information on how to call the <code>PutItem</code> API using the AWS SDK in specific languages, see the following:</p> <ul> <li> <p> <a href="http://docs.aws.amazon.com/goto/aws-cli/dynamodb-2012-08-10/PutItem"> PutItem in the AWS Command Line Interface </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/DotNetSDKV3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for .NET </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForCpp/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for C++ </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForGoV1/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Go </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForJava/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Java </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/AWSJavaScriptSDK/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for JavaScript </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForPHPV3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for PHP V3 </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/boto3/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Python </a> </p> </li> <li> <p> <a href="http://docs.aws.amazon.com/goto/SdkForRubyV2/dynamodb-2012-08-10/PutItem"> PutItem in the AWS SDK for Ruby V2 </a> </p> </li> </ul> </important> <p>When you add an item, the primary key attribute(s) are the only required attributes. Attribute values cannot be null. String and Binary type attributes must have lengths greater than zero. Set type attributes cannot be empty. Requests with empty values will be rejected with a <code>ValidationException</code> exception.</p> <note> <p>To prevent a new item from replacing an existing item, use a conditional expression that contains the <code>attribute_not_exists</code> function with the name of the attribute being used as the partition key for the table. Since every record must contain that attribute, the <code>attribute_not_exists</code> function will only succeed if no matching item exists.</p> </note> <p>For more information about <code>PutItem</code>, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/WorkingWithItems.html">Working with Items</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn put_item(&self, input: PutItemInput) -> RusotoFuture<PutItemOutput, PutItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.PutItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response).deserialize::<PutItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(PutItemError::from_response(response))), ) } }) } /// <p>The <code>Query</code> operation finds items based on primary key values. You can query any table or secondary index that has a composite primary key (a partition key and a sort key). </p> <p>Use the <code>KeyConditionExpression</code> parameter to provide a specific value for the partition key. The <code>Query</code> operation will return all of the items from the table or index with that partition key value. You can optionally narrow the scope of the <code>Query</code> operation by specifying a sort key value and a comparison operator in <code>KeyConditionExpression</code>. To further refine the <code>Query</code> results, you can optionally provide a <code>FilterExpression</code>. A <code>FilterExpression</code> determines which items within the results should be returned to you. All of the other results are discarded. </p> <p> A <code>Query</code> operation always returns a result set. If no matching items are found, the result set will be empty. Queries that do not return results consume the minimum number of read capacity units for that type of read operation. </p> <note> <p> DynamoDB calculates the number of read capacity units consumed based on item size, not on the amount of data that is returned to an application. The number of capacity units consumed will be the same whether you request all of the attributes (the default behavior) or just some of them (using a projection expression). The number will also be the same whether or not you use a <code>FilterExpression</code>. </p> </note> <p> <code>Query</code> results are always sorted by the sort key value. If the data type of the sort key is Number, the results are returned in numeric order; otherwise, the results are returned in order of UTF-8 bytes. By default, the sort order is ascending. To reverse the order, set the <code>ScanIndexForward</code> parameter to false. </p> <p> A single <code>Query</code> operation will read up to the maximum number of items set (if using the <code>Limit</code> parameter) or a maximum of 1 MB of data and then apply any filtering to the results using <code>FilterExpression</code>. If <code>LastEvaluatedKey</code> is present in the response, you will need to paginate the result set. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Query.html#Query.Pagination">Paginating the Results</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> <p> <code>FilterExpression</code> is applied after a <code>Query</code> finishes, but before the results are returned. A <code>FilterExpression</code> cannot contain partition key or sort key attributes. You need to specify those attributes in the <code>KeyConditionExpression</code>. </p> <note> <p> A <code>Query</code> operation can return an empty result set and a <code>LastEvaluatedKey</code> if all the items read for the page of results are filtered out. </p> </note> <p>You can query a table, a local secondary index, or a global secondary index. For a query on a table or on a local secondary index, you can set the <code>ConsistentRead</code> parameter to <code>true</code> and obtain a strongly consistent result. Global secondary indexes support eventually consistent reads only, so do not specify <code>ConsistentRead</code> when querying a global secondary index.</p> fn query(&self, input: QueryInput) -> RusotoFuture<QueryOutput, QueryError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.Query"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response).deserialize::<QueryOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(QueryError::from_response(response))), ) } }) } /// <p><p>Creates a new table from an existing backup. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. </p> <p>You can call <code>RestoreTableFromBackup</code> at a maximum rate of 10 times per second.</p> <p>You must manually set up the following on the restored table:</p> <ul> <li> <p>Auto scaling policies</p> </li> <li> <p>IAM policies</p> </li> <li> <p>Cloudwatch metrics and alarms</p> </li> <li> <p>Tags</p> </li> <li> <p>Stream settings</p> </li> <li> <p>Time to Live (TTL) settings</p> </li> </ul></p> fn restore_table_from_backup( &self, input: RestoreTableFromBackupInput, ) -> RusotoFuture<RestoreTableFromBackupOutput, RestoreTableFromBackupError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.RestoreTableFromBackup"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<RestoreTableFromBackupOutput, _>() })) } else { Box::new( response.buffer().from_err().and_then(|response| { Err(RestoreTableFromBackupError::from_response(response)) }), ) } }) } /// <p><p>Restores the specified table to the specified point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. You can restore your table to any point in time during the last 35 days. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account. </p> <p> When you restore using point in time recovery, DynamoDB restores your table data to the state based on the selected date and time (day:hour:minute:second) to a new table. </p> <p> Along with data, the following are also included on the new restored table using point in time recovery: </p> <ul> <li> <p>Global secondary indexes (GSIs)</p> </li> <li> <p>Local secondary indexes (LSIs)</p> </li> <li> <p>Provisioned read and write capacity</p> </li> <li> <p>Encryption settings</p> <important> <p> All these settings come from the current settings of the source table at the time of restore. </p> </important> </li> </ul> <p>You must manually set up the following on the restored table:</p> <ul> <li> <p>Auto scaling policies</p> </li> <li> <p>IAM policies</p> </li> <li> <p>Cloudwatch metrics and alarms</p> </li> <li> <p>Tags</p> </li> <li> <p>Stream settings</p> </li> <li> <p>Time to Live (TTL) settings</p> </li> <li> <p>Point in time recovery settings</p> </li> </ul></p> fn restore_table_to_point_in_time( &self, input: RestoreTableToPointInTimeInput, ) -> RusotoFuture<RestoreTableToPointInTimeOutput, RestoreTableToPointInTimeError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header( "x-amz-target", "DynamoDB_20120810.RestoreTableToPointInTime", ); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<RestoreTableToPointInTimeOutput, _>() })) } else { Box::new(response.buffer().from_err().and_then(|response| { Err(RestoreTableToPointInTimeError::from_response(response)) })) } }) } /// <p>The <code>Scan</code> operation returns one or more items and item attributes by accessing every item in a table or a secondary index. To have DynamoDB return fewer items, you can provide a <code>FilterExpression</code> operation.</p> <p>If the total number of scanned items exceeds the maximum data set size limit of 1 MB, the scan stops and results are returned to the user as a <code>LastEvaluatedKey</code> value to continue the scan in a subsequent operation. The results also include the number of items exceeding the limit. A scan can result in no table data meeting the filter criteria. </p> <p>A single <code>Scan</code> operation will read up to the maximum number of items set (if using the <code>Limit</code> parameter) or a maximum of 1 MB of data and then apply any filtering to the results using <code>FilterExpression</code>. If <code>LastEvaluatedKey</code> is present in the response, you will need to paginate the result set. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Scan.html#Scan.Pagination">Paginating the Results</a> in the <i>Amazon DynamoDB Developer Guide</i>. </p> <p> <code>Scan</code> operations proceed sequentially; however, for faster performance on a large table or secondary index, applications can request a parallel <code>Scan</code> operation by providing the <code>Segment</code> and <code>TotalSegments</code> parameters. For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Scan.html#Scan.ParallelScan">Parallel Scan</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> <p> <code>Scan</code> uses eventually consistent reads when accessing the data in a table; therefore, the result set might not include the changes to data in the table immediately before the operation began. If you need a consistent copy of the data, as of the time that the <code>Scan</code> begins, you can set the <code>ConsistentRead</code> parameter to <code>true</code>.</p> fn scan(&self, input: ScanInput) -> RusotoFuture<ScanOutput, ScanError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.Scan"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response).deserialize::<ScanOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(ScanError::from_response(response))), ) } }) } /// <p>Associate a set of tags with an Amazon DynamoDB resource. You can then activate these user-defined tags so that they appear on the Billing and Cost Management console for cost allocation tracking. You can call TagResource up to 5 times per second, per account. </p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn tag_resource(&self, input: TagResourceInput) -> RusotoFuture<(), TagResourceError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.TagResource"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(future::ok(::std::mem::drop(response))) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(TagResourceError::from_response(response))), ) } }) } /// <p><p> <code>TransactGetItems</code> is a synchronous operation that atomically retrieves multiple items from one or more tables (but not from indexes) in a single account and region. A <code>TransactGetItems</code> call can contain up to 10 <code>TransactGetItem</code> objects, each of which contains a <code>Get</code> structure that specifies an item to retrieve from a table in the account and region. A call to <code>TransactGetItems</code> cannot retrieve items from tables in more than one AWS account or region.</p> <p>DynamoDB rejects the entire <code>TransactGetItems</code> request if any of the following is true:</p> <ul> <li> <p>A conflicting operation is in the process of updating an item to be read.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul></p> fn transact_get_items( &self, input: TransactGetItemsInput, ) -> RusotoFuture<TransactGetItemsOutput, TransactGetItemsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.TransactGetItems"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<TransactGetItemsOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(TransactGetItemsError::from_response(response))), ) } }) } /// <p><p> <code>TransactWriteItems</code> is a synchronous write operation that groups up to 10 action requests. These actions can target items in different tables, but not in different AWS accounts or regions, and no two actions can target the same item. For example, you cannot both <code>ConditionCheck</code> and <code>Update</code> the same item.</p> <p>The actions are completed atomically so that either all of them succeed, or all of them fail. They are defined by the following objects:</p> <ul> <li> <p> <code>Put</code> — Initiates a <code>PutItem</code> operation to write a new item. This structure specifies the primary key of the item to be written, the name of the table to write it in, an optional condition expression that must be satisfied for the write to succeed, a list of the item's attributes, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>Update</code> — Initiates an <code>UpdateItem</code> operation to update an existing item. This structure specifies the primary key of the item to be updated, the name of the table where it resides, an optional condition expression that must be satisfied for the update to succeed, an expression that defines one or more attributes to be updated, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>Delete</code> — Initiates a <code>DeleteItem</code> operation to delete an existing item. This structure specifies the primary key of the item to be deleted, the name of the table where it resides, an optional condition expression that must be satisfied for the deletion to succeed, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> <li> <p> <code>ConditionCheck</code> — Applies a condition to an item that is not being modified by the transaction. This structure specifies the primary key of the item to be checked, the name of the table where it resides, a condition expression that must be satisfied for the transaction to succeed, and a field indicating whether or not to retrieve the item's attributes if the condition is not met.</p> </li> </ul> <p>DynamoDB rejects the entire <code>TransactWriteItems</code> request if any of the following is true:</p> <ul> <li> <p>A condition in one of the condition expressions is not met.</p> </li> <li> <p>A conflicting operation is in the process of updating the same item.</p> </li> <li> <p>There is insufficient provisioned capacity for the transaction to be completed.</p> </li> <li> <p>An item size becomes too large (bigger than 400 KB), a Local Secondary Index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.</p> </li> <li> <p>There is a user error, such as an invalid data format.</p> </li> </ul></p> fn transact_write_items( &self, input: TransactWriteItemsInput, ) -> RusotoFuture<TransactWriteItemsOutput, TransactWriteItemsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.TransactWriteItems"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<TransactWriteItemsOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(TransactWriteItemsError::from_response(response))), ) } }) } /// <p>Removes the association of tags from an Amazon DynamoDB resource. You can call UntagResource up to 5 times per second, per account. </p> <p>For an overview on tagging DynamoDB resources, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Tagging.html">Tagging for DynamoDB</a> in the <i>Amazon DynamoDB Developer Guide</i>.</p> fn untag_resource(&self, input: UntagResourceInput) -> RusotoFuture<(), UntagResourceError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UntagResource"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(future::ok(::std::mem::drop(response))) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(UntagResourceError::from_response(response))), ) } }) } /// <p> <code>UpdateContinuousBackups</code> enables or disables point in time recovery for the specified table. A successful <code>UpdateContinuousBackups</code> call returns the current <code>ContinuousBackupsDescription</code>. Continuous backups are <code>ENABLED</code> on all tables at table creation. If point in time recovery is enabled, <code>PointInTimeRecoveryStatus</code> will be set to ENABLED.</p> <p> Once continuous backups and point in time recovery are enabled, you can restore to any point in time within <code>EarliestRestorableDateTime</code> and <code>LatestRestorableDateTime</code>. </p> <p> <code>LatestRestorableDateTime</code> is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days.. </p> fn update_continuous_backups( &self, input: UpdateContinuousBackupsInput, ) -> RusotoFuture<UpdateContinuousBackupsOutput, UpdateContinuousBackupsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UpdateContinuousBackups"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateContinuousBackupsOutput, _>() })) } else { Box::new(response.buffer().from_err().and_then(|response| { Err(UpdateContinuousBackupsError::from_response(response)) })) } }) } /// <p><p>Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, must have the same name as the global table, must have the same key schema, and must have DynamoDB Streams enabled and must have same provisioned and maximum write capacity units.</p> <note> <p>Although you can use <code>UpdateGlobalTable</code> to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas.</p> </note> <p> If global secondary indexes are specified, then the following conditions must also be met: </p> <ul> <li> <p> The global secondary indexes must have the same name. </p> </li> <li> <p> The global secondary indexes must have the same hash key and sort key (if present). </p> </li> <li> <p> The global secondary indexes must have the same provisioned and maximum write capacity units. </p> </li> </ul></p> fn update_global_table( &self, input: UpdateGlobalTableInput, ) -> RusotoFuture<UpdateGlobalTableOutput, UpdateGlobalTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UpdateGlobalTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateGlobalTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(UpdateGlobalTableError::from_response(response))), ) } }) } /// <p>Updates settings for a global table.</p> fn update_global_table_settings( &self, input: UpdateGlobalTableSettingsInput, ) -> RusotoFuture<UpdateGlobalTableSettingsOutput, UpdateGlobalTableSettingsError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header( "x-amz-target", "DynamoDB_20120810.UpdateGlobalTableSettings", ); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateGlobalTableSettingsOutput, _>() })) } else { Box::new(response.buffer().from_err().and_then(|response| { Err(UpdateGlobalTableSettingsError::from_response(response)) })) } }) } /// <p>Edits an existing item's attributes, or adds a new item to the table if it does not already exist. You can put, delete, or add attribute values. You can also perform a conditional update on an existing item (insert a new attribute name-value pair if it doesn't exist, or replace an existing name-value pair if it has certain expected attribute values).</p> <p>You can also return the item's attribute values in the same <code>UpdateItem</code> operation using the <code>ReturnValues</code> parameter.</p> fn update_item( &self, input: UpdateItemInput, ) -> RusotoFuture<UpdateItemOutput, UpdateItemError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UpdateItem"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateItemOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(UpdateItemError::from_response(response))), ) } }) } /// <p>Modifies the provisioned throughput settings, global secondary indexes, or DynamoDB Streams settings for a given table.</p> <p>You can only perform one of the following operations at once:</p> <ul> <li> <p>Modify the provisioned throughput settings of the table.</p> </li> <li> <p>Enable or disable Streams on the table.</p> </li> <li> <p>Remove a global secondary index from the table.</p> </li> <li> <p>Create a new global secondary index on the table. Once the index begins backfilling, you can use <code>UpdateTable</code> to perform other operations.</p> </li> </ul> <p> <code>UpdateTable</code> is an asynchronous operation; while it is executing, the table status changes from <code>ACTIVE</code> to <code>UPDATING</code>. While it is <code>UPDATING</code>, you cannot issue another <code>UpdateTable</code> request. When the table returns to the <code>ACTIVE</code> state, the <code>UpdateTable</code> operation is complete.</p> fn update_table( &self, input: UpdateTableInput, ) -> RusotoFuture<UpdateTableOutput, UpdateTableError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UpdateTable"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateTableOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(UpdateTableError::from_response(response))), ) } }) } /// <p>The UpdateTimeToLive method will enable or disable TTL for the specified table. A successful <code>UpdateTimeToLive</code> call returns the current <code>TimeToLiveSpecification</code>; it may take up to one hour for the change to fully process. Any additional <code>UpdateTimeToLive</code> calls for the same table during this one hour duration result in a <code>ValidationException</code>. </p> <p>TTL compares the current time in epoch time format to the time stored in the TTL attribute of an item. If the epoch time value stored in the attribute is less than the current time, the item is marked as expired and subsequently deleted.</p> <note> <p> The epoch time format is the number of seconds elapsed since 12:00:00 AM January 1st, 1970 UTC. </p> </note> <p>DynamoDB deletes expired items on a best-effort basis to ensure availability of throughput for other data operations. </p> <important> <p>DynamoDB typically deletes expired items within two days of expiration. The exact duration within which an item gets deleted after expiration is specific to the nature of the workload. Items that have expired and not been deleted will still show up in reads, queries, and scans.</p> </important> <p>As items are deleted, they are removed from any Local Secondary Index and Global Secondary Index immediately in the same eventually consistent way as a standard delete operation.</p> <p>For more information, see <a href="https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/TTL.html">Time To Live</a> in the Amazon DynamoDB Developer Guide. </p> fn update_time_to_live( &self, input: UpdateTimeToLiveInput, ) -> RusotoFuture<UpdateTimeToLiveOutput, UpdateTimeToLiveError> { let mut request = SignedRequest::new("POST", "dynamodb", &self.region, "/"); request.set_content_type("application/x-amz-json-1.0".to_owned()); request.add_header("x-amz-target", "DynamoDB_20120810.UpdateTimeToLive"); let encoded = serde_json::to_string(&input).unwrap(); request.set_payload(Some(encoded)); self.client.sign_and_dispatch(request, |response| { if response.status.is_success() { Box::new(response.buffer().from_err().and_then(|response| { proto::json::ResponsePayload::new(&response) .deserialize::<UpdateTimeToLiveOutput, _>() })) } else { Box::new( response .buffer() .from_err() .and_then(|response| Err(UpdateTimeToLiveError::from_response(response))), ) } }) } }