Modular and SF Compute: Up to 80% lower cost per token 🚀 @media only screen and (max-width:639px){img.stretch-on-mobile,.hs_rss_email_entries_table img,.hs-stretch-cta .hs-cta-img{height:auto !important;width:100% !important} .display_block_on_small_screens{display:block}.hs_padded{padding-left:20px !important;padding-right:20px !important} .hs-hm,table.hs-hm{display:none}.hs-hd{display:block !important}table.hs-hd{display:table !important} }@media only screen and (max-width:639px){.hse-border-m{border-left:1px solid #cbd6e2 !important;border-right:1px solid #cbd6e2 !important;box-sizing:border-box} .hse-border-bottom-m{border-bottom:1px solid #cbd6e2 !important}.hse-border-top-m{border-top:1px solid #cbd6e2 !important} .hse-border-top-hm{border-top:none !important}.hse-border-bottom-hm{border-bottom:none !important} }.moz-text-html .hse-column-container{max-width:600px !important;width:600px !important} .moz-text-html .hse-column{display:table-cell;vertical-align:top}.moz-text-html .hse-section .hse-size-6{max-width:300px !important;width:300px !important} .moz-text-html .hse-section .hse-size-12{max-width:600px !important;width:600px !important} @media only screen and (min-width:640px){.hse-column-container{max-width:600px !important;width:600px !important} .hse-column{display:table-cell;vertical-align:top}.hse-section .hse-size-6{max-width:300px !important;width:300px !important} .hse-section .hse-size-12{max-width:600px !important;width:600px !important} }@media only screen and (max-width:639px){.hse-body-wrapper-td{padding-top:20px !important} #section-0 .hse-column-container{padding-top:10px !important;padding-bottom:10px !important;background-color:transparent !important} #section-0 .hse-column-container{background-color:transparent !important} }@media only screen and (max-width:639px){ #section-1 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-1 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-2 .hse-column-container{padding-top:10px !important;padding-bottom:0px !important} #section-2 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-3 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-3 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-4 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-4 .hse-column-container{background-color:#fff !important} }@media screen and (max-width:639px){.social-network-cell{display:inline-block} }@media only screen and (max-width:639px){ #section-6 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-6 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-7 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-7 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){.hse-body-wrapper-td{padding-bottom:20px !important} #section-8 .hse-column-container{padding-top:30px !important;padding-bottom:0px !important;background-color:transparent !important} #section-8 .hse-column-container{background-color:transparent !important} }#hs_body #hs_cos_wrapper_main a[x-apple-data-detectors]{color:inherit !important;text-decoration:none !important;font-size:inherit !important;font-family:inherit !important;font-weight:inherit !important;line-height:inherit !important} a{text-decoration:underline}p{margin:0}body{-ms-text-size-adjust:100%;-webkit-text-size-adjust:100%;-webkit-font-smoothing:antialiased;moz-osx-font-smoothing:grayscale} table{border-spacing:0;mso-table-lspace:0;mso-table-rspace:0}table,td{border-collapse:collapse} img{-ms-interpolation-mode:bicubic}p,a,li,td,blockquote{mso-line-height-rule:exactly}
We’re launching a new API that delivers large-scale inference at up to 80% lower cost. See the demo and learn more.
We’ve partnered with SF Compute to launch the Large Scale Inference Batch API, a new way to run large-scale inference at a fraction of today’s cost.
Built for high-throughput, asynchronous workloads, this new API combines Modular’s ultra-efficient inference stack with SF Compute’s real-time GPU spot market. The result:
Whether you’re running billions of tokens or images, this stack is designed to scale efficiently without breaking your budget. Try it today - we’re offering 10M’s of batch inference tokens for FREE to the first 100 new customers!
Modular and SF Compute: Up to 80% lower cost per token 🚀 @media only screen and (max-width:639px){img.stretch-on-mobile,.hs_rss_email_entries_table img,.hs-stretch-cta .hs-cta-img{height:auto !important;width:100% !important} .display_block_on_small_screens{display:block}.hs_padded{padding-left:20px !important;padding-right:20px !important} .hs-hm,table.hs-hm{display:none}.hs-hd{display:block !important}table.hs-hd{display:table !important} }@media only screen and (max-width:639px){.hse-border-m{border-left:1px solid #cbd6e2 !important;border-right:1px solid #cbd6e2 !important;box-sizing:border-box} .hse-border-bottom-m{border-bottom:1px solid #cbd6e2 !important}.hse-border-top-m{border-top:1px solid #cbd6e2 !important} .hse-border-top-hm{border-top:none !important}.hse-border-bottom-hm{border-bottom:none !important} }.moz-text-html .hse-column-container{max-width:600px !important;width:600px !important} .moz-text-html .hse-column{display:table-cell;vertical-align:top}.moz-text-html .hse-section .hse-size-6{max-width:300px !important;width:300px !important} .moz-text-html .hse-section .hse-size-12{max-width:600px !important;width:600px !important} @media only screen and (min-width:640px){.hse-column-container{max-width:600px !important;width:600px !important} .hse-column{display:table-cell;vertical-align:top}.hse-section .hse-size-6{max-width:300px !important;width:300px !important} .hse-section .hse-size-12{max-width:600px !important;width:600px !important} }@media only screen and (max-width:639px){.hse-body-wrapper-td{padding-top:20px !important} #section-0 .hse-column-container{padding-top:10px !important;padding-bottom:10px !important;background-color:transparent !important} #section-0 .hse-column-container{background-color:transparent !important} }@media only screen and (max-width:639px){ #section-1 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-1 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-2 .hse-column-container{padding-top:10px !important;padding-bottom:0px !important} #section-2 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-3 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-3 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-4 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-4 .hse-column-container{background-color:#fff !important} }@media screen and (max-width:639px){.social-network-cell{display:inline-block} }@media only screen and (max-width:639px){ #section-6 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-6 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){ #section-7 .hse-column-container{padding-top:0px !important;padding-bottom:0px !important} #section-7 .hse-column-container{background-color:#fff !important} }@media only screen and (max-width:639px){.hse-body-wrapper-td{padding-bottom:20px !important} #section-8 .hse-column-container{padding-top:30px !important;padding-bottom:0px !important;background-color:transparent !important} #section-8 .hse-column-container{background-color:transparent !important} }#hs_body #hs_cos_wrapper_main a[x-apple-data-detectors]{color:inherit !important;text-decoration:none !important;font-size:inherit !important;font-family:inherit !important;font-weight:inherit !important;line-height:inherit !important} a{text-decoration:underline}p{margin:0}body{-ms-text-size-adjust:100%;-webkit-text-size-adjust:100%;-webkit-font-smoothing:antialiased;moz-osx-font-smoothing:grayscale} table{border-spacing:0;mso-table-lspace:0;mso-table-rspace:0}table,td{border-collapse:collapse} img{-ms-interpolation-mode:bicubic}p,a,li,td,blockquote{mso-line-height-rule:exactly}
We’re launching a new API that delivers large-scale inference at up to 80% lower cost. See the demo and learn more.
We’ve partnered with SF Compute to launch the Large Scale Inference Batch API, a new way to run large-scale inference at a fraction of today’s cost.
Built for high-throughput, asynchronous workloads, this new API combines Modular’s ultra-efficient inference stack with SF Compute’s real-time GPU spot market. The result:
Whether you’re running billions of tokens or images, this stack is designed to scale efficiently without breaking your budget. Try it today - we’re offering 10M’s of batch inference tokens for FREE to the first 100 new customers!
发布者