Scala

[ follow ]
Scala
fromScala-lang
1 week ago

The Sovereign Tech Fund invests in Scala

Sovereign Tech Fund invested €377,300 in Scala via the Scala Center for a two-year program to strengthen security, maintenance, and developer experience.
Scala
fromMedium
1 week ago

I Thought Scala Was Vibe Coding

Scala emphasizes immutability, expression-oriented programming, powerful pattern matching, and Option-based null safety for more concise, safer, and more composable JVM code.
fromMedium
1 week ago

From Scala to Kotlin: A Frustrated Engineer's Honest Take

Over the past months, I've watched two clients move from Scala (Play, Slick, Akka, Akka Http ... ) to Kotlin (Spring, JPA/Hibernate). In my current role, an engineering decision was made to move away from Scala. The decision was driven less by Scala's shortcomings and more by long-term career risk management: leaders understandably favor stacks (Java/Kotlin) that maximize hiring flexibility in a volatile market.
Scala
Scala
fromScala-lang
1 week ago

Scala 3.8 released!

Scala 3.8 modernizes the ecosystem, requires JDK 17+, stabilizes key features, offers experimental features, and users should use 3.8.1 due to a runtime regression.
fromMedium
2 weeks ago

Agentic Workflows in Scala (Without the Buzzwords)

High-level view of the travel search workflow, highlighting parallel searches, explicit decision points, and iterative refinement. In Scala, we define this workflow using Workflows4s, encoding both state and transitions explicitly in the type system. Instead of opaque state blobs or untyped contexts, the state of the process is represented using algebraic data types - types like Started, Found, Sent, and Booked - each corresponding to a distinct point in the workflow's lifecycle.
Scala
Scala
fromMedium
1 month ago

Scala Programming Explained: A Complete Storytelling Guide for Students and Developers

Scala blends object-oriented and functional programming on the JVM to deliver scalable, concise, high-performance solutions for backend, big data, and enterprise systems.
Scala
fromMedium
1 month ago

Scala Dependency Injection for Java Developers (Part 1): Why Spring DI Feels Heavy

Spring DI uses runtime reflection and implicit wiring, giving convenience but causing hidden coupling, runtime resolution, and added complexity compared with Scala's explicit, compile-time approach.
#scala-3
fromMedium
1 month ago
Scala

Scala Dependency Injection for Java Developers (Part 2): Given and Using Explained

fromMedium
1 month ago
Scala

Scala Dependency Injection for Java Developers (Part 2): Given and Using Explained

Scala
fromMedium
1 month ago

Why Metaprogramming Usually Makes Things Worse

Metaprogramming tools like macros and templates are rarely necessary and often harmful, increasing compilation costs, complexity, and slowing development feedback loops.
fromMedium
1 month ago

Type Classes: How Scala Achieves Polymorphism Without Inheritance

Imagine you're working with a third-party library that provides a User class. You need to add JSON serialization to it, but you can't modify the source code. Of course you can create a wrapper class or extend it, but that feels clunky and breaks existing code that expects the original type. This is where type classes shine. They're one of Scala's most powerful patterns, and they're the secret ingredient in popular libraries like Cats, Scalaz, and Circe.
Scala
fromMedium
1 month ago

Snowpark Scala 2.13 Public Preview

Photo by Goran Ivos on Unsplash We're excited to announce that Scala 2.13 is now in Public Preview (PuPr) on Snowpark for Scala client, UDxF and Stored Procedures! This release brings the massive collections overhaul, performance improvements, and powerful language enhancements of Scala 2.13 to the Snowflake AI Data Cloud. Where Can You Use Scala 2.13 in Snowflake? Snowflake SQL Snowpark Scala client library Why Upgrade?
Scala
fromMedium
2 months ago

TPC-DS on Spark 4.0+ : A Practical Guide and Benchmarking Considerations

The spark-sql-perf toolkit doesn't work for Spark 4.0+ currently, and this guide shows you how to get it running(with a custom patch). While many developers have their own complex Spark setup, this workflow is designed to be simple and reproducible. It only requires an AWS account to provision a cluster and run a full benchmark from scratch. We'll focus on the patch, the build process, and how a tool like FlintRock makes deploying custom Spark clusters incredibly simple.
Scala
#spark
[ Load more ]