From b54a6df86997568891edff4af0dd9cb2e1e799f6 Mon Sep 17 00:00:00 2001 From: Enno Runne <458526+ennru@users.noreply.github.com> Date: Thu, 10 Oct 2024 16:58:35 +0200 Subject: [PATCH] docs: revise URL structure (libraries) (#1773) --- .github/workflows/link-validator.yml | 4 +-- README.md | 8 +++--- build.sbt | 26 ++++++++----------- core/src/main/resources/reference.conf | 6 ++--- .../internal/ConsumerResetProtection.scala | 4 +-- .../scala/akka/kafka/javadsl/Consumer.scala | 4 +-- .../scala/akka/kafka/javadsl/Producer.scala | 4 +-- .../akka/kafka/javadsl/Transactional.scala | 4 +-- .../scala/akka/kafka/scaladsl/Consumer.scala | 12 ++++----- .../kafka/scaladsl/DiscoverySupport.scala | 2 +- .../scala/akka/kafka/scaladsl/Producer.scala | 4 +-- .../akka/kafka/scaladsl/Transactional.scala | 4 +-- docs/release-train-issue-template.md | 8 +++--- docs/src/main/paradox/.htaccess | 2 +- docs/src/main/paradox/home.md | 6 ++--- docs/src/main/paradox/index.md | 2 +- docs/src/main/paradox/release-notes/1.0.x.md | 18 ++++++------- docs/src/main/paradox/release-notes/1.1.x.md | 4 +-- docs/src/main/paradox/release-notes/2.0.x.md | 2 +- docs/src/main/paradox/serialization.md | 2 +- docs/src/main/paradox/snapshots.md | 2 +- .../main/paradox/testing-testcontainers.md | 4 +-- project/project-info.conf | 2 +- scripts/link-validator.conf | 8 +++--- .../TestcontainersKafkaJunit4Test.java | 2 +- .../javadsl/TestcontainersKafkaTest.java | 2 +- .../scaladsl/TestcontainersKafkaLike.scala | 2 +- .../TestcontainersKafkaPerClassLike.scala | 2 +- .../akka/kafka/scaladsl/IntegrationSpec.scala | 2 +- .../akka/kafka/tests/CapturingAppender.scala | 4 +-- .../scala/akka/kafka/tests/LogbackUtil.scala | 2 +- .../tests/javadsl/LogCapturingJunit4.scala | 2 +- .../kafka/tests/scaladsl/LogCapturing.scala | 2 +- 33 files changed, 80 insertions(+), 82 deletions(-) diff --git a/.github/workflows/link-validator.yml b/.github/workflows/link-validator.yml index d296b83a0..ac37576de 100644 --- a/.github/workflows/link-validator.yml +++ b/.github/workflows/link-validator.yml @@ -4,7 +4,7 @@ on: workflow_dispatch: pull_request: schedule: - - cron: '0 6 * * 1' + - cron: '10 6 1 * *' permissions: contents: read # allow actions/checkout @@ -43,4 +43,4 @@ jobs: run: sbt docs/makeSite - name: Run Link Validator - run: cs launch net.runne::site-link-validator:0.2.3 -- scripts/link-validator.conf + run: cs launch net.runne::site-link-validator:0.2.5 -- scripts/link-validator.conf diff --git a/README.md b/README.md index 201c71fab..c391c9bb0 100644 --- a/README.md +++ b/README.md @@ -7,7 +7,7 @@ Alpakka Kafka [![gh-actions-badge][]][gh-actions] Systems don't come alone. In the modern world of microservices and cloud deployment, new components must interact with legacy systems, making integration an important key to success. Reactive Streams give us a technology-independent tool to let these heterogeneous systems communicate without overwhelming each other. -The Alpakka project is an open source initiative to implement stream-aware, reactive, integration pipelines for Java and Scala. It is built on top of [Akka Streams](https://doc.akka.io/docs/akka/current/stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK 9+ [java.util.concurrent.Flow](https://docs.oracle.com/javase/10/docs/api/java/util/concurrent/Flow.html)-compliant implementation and therefore [fully interoperable](https://doc.akka.io/docs/akka/current/general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. +The Alpakka project is an open source initiative to implement stream-aware, reactive, integration pipelines for Java and Scala. It is built on top of [Akka Streams](https://doc.akka.io/libraries/akka-core/current/stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK 9+ [java.util.concurrent.Flow](https://docs.oracle.com/javase/10/docs/api/java/util/concurrent/Flow.html)-compliant implementation and therefore [fully interoperable](https://doc.akka.io/libraries/akka-core/current/general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. This repository contains the sources for the **Alpakka Kafka connector**. Which lets you connect [Apache Kafka](https://kafka.apache.org/) to Akka Streams. It was formerly known as **Akka Streams Kafka** and even **Reactive Kafka**. @@ -17,9 +17,9 @@ Akka Stream connectors to other technologies are listed in the [Alpakka reposito Documentation ------------- -- [Alpakka reference](https://doc.akka.io/docs/alpakka/current/) documentation +- [Alpakka reference](https://doc.akka.io/libraries/alpakka/current/) documentation -- **[Alpakka Kafka connector reference](https://doc.akka.io/docs/akka-stream-kafka/current/) documentation** +- **[Alpakka Kafka connector reference](https://doc.akka.io/libraries/alpakka-kafka/current/) documentation** To keep up with the latest Alpakka releases check out [Alpakka releases](https://github.com/akka/alpakka/releases) and [Alpakka Kafka releases](https://github.com/akka/alpakka-kafka/releases). @@ -67,4 +67,4 @@ License Akka is licensed under the Business Source License 1.1, please see the [Akka License FAQ](https://www.lightbend.com/akka/license-faq). -Tests and documentation are under a separate license, see the LICENSE file in each documentation and test root directory for details. \ No newline at end of file +Tests and documentation are under a separate license, see the LICENSE file in each documentation and test root directory for details. diff --git a/build.sbt b/build.sbt index 2b515d18d..9ccb11b4e 100644 --- a/build.sbt +++ b/build.sbt @@ -75,7 +75,7 @@ val commonSettings = Def.settings( organization := "com.typesafe.akka", organizationName := "Lightbend Inc.", organizationHomepage := Some(url("https://www.lightbend.com/")), - homepage := Some(url("https://doc.akka.io/docs/alpakka-kafka/current")), + homepage := Some(url("https://doc.akka.io/libraries/alpakka-kafka/current")), scmInfo := Some(ScmInfo(url("https://github.com/akka/alpakka-kafka"), "git@github.com:akka/alpakka-kafka.git")), developers += Developer("contributors", "Contributors", @@ -129,19 +129,15 @@ val commonSettings = Def.settings( "https://doc.akka.io/api/alpakka-kafka/current/" ) ++ { if (scalaBinaryVersion.value.startsWith("3")) { - Seq("-skip-packages:akka.pattern") // different usage in scala3 + Seq(s"-external-mappings:https://docs.oracle.com/en/java/javase/${JavaDocLinkVersion}/docs/api/java.base/", + "-skip-packages:akka.pattern") } else { - Seq("-skip-packages", "akka.pattern") // for some reason Scaladoc creates this + Seq("-jdk-api-doc-base", + s"https://docs.oracle.com/en/java/javase/${JavaDocLinkVersion}/docs/api/java.base/", + "-skip-packages", + "akka.pattern") } }, - // make use of https://github.com/scala/scala/pull/8663 - Compile / doc / scalacOptions ++= { - if (scalaBinaryVersion.value.startsWith("3")) { - Seq(s"-external-mappings:https://docs.oracle.com/en/java/javase/${JavaDocLinkVersion}/docs/api/java.base/") // different usage in scala3 - } else if (scalaBinaryVersion.value.startsWith("2.13")) { - Seq("-jdk-api-doc-base", s"https://docs.oracle.com/en/java/javase/${JavaDocLinkVersion}/docs/api/java.base/") - } else Nil - }, Compile / doc / scalacOptions -= "-Xfatal-warnings", // show full stack traces and test case durations testOptions += Tests.Argument(TestFrameworks.ScalaTest, "-oDF"), @@ -185,7 +181,7 @@ lazy val `alpakka-kafka` = | testkit - framework for testing the connector | |Other modules: - | docs - the sources for generating https://doc.akka.io/docs/alpakka-kafka/current + | docs - the sources for generating https://doc.akka.io/libraries/alpakka-kafka/current | benchmarks - compare direct Kafka API usage with Alpakka Kafka | |Useful sbt tasks: @@ -342,7 +338,7 @@ lazy val docs = project Preprocess / preprocessRules := Seq( ("https://javadoc\\.io/page/".r, _ => "https://javadoc\\.io/static/") ), - Paradox / siteSubdirName := s"docs/alpakka-kafka/${projectInfoVersion.value}", + Paradox / siteSubdirName := s"libraries/alpakka-kafka/${projectInfoVersion.value}", paradoxGroups := Map("Language" -> Seq("Java", "Scala")), paradoxProperties ++= Map( "image.base_url" -> "images/", @@ -352,11 +348,11 @@ lazy val docs = project "javadoc.akka.kafka.base_url" -> "", // Akka "akka.version" -> akkaVersion, - "extref.akka.base_url" -> s"https://doc.akka.io/docs/akka/$AkkaBinaryVersionForDocs/%s", + "extref.akka.base_url" -> s"https://doc.akka.io/libraries/akka-core/$AkkaBinaryVersionForDocs/%s", "scaladoc.akka.base_url" -> s"https://doc.akka.io/api/akka/$AkkaBinaryVersionForDocs/", "javadoc.akka.base_url" -> s"https://doc.akka.io/japi/akka/$AkkaBinaryVersionForDocs/", "javadoc.akka.link_style" -> "direct", - "extref.akka-management.base_url" -> s"https://doc.akka.io/docs/akka-management/current/%s", + "extref.akka-management.base_url" -> s"https://doc.akka.io/libraries/akka-management/current/%s", // Kafka "kafka.version" -> kafkaVersion, "extref.kafka.base_url" -> s"https://kafka.apache.org/$KafkaVersionForDocs/%s", diff --git a/core/src/main/resources/reference.conf b/core/src/main/resources/reference.conf index 2ebd19d3f..368a83dcb 100644 --- a/core/src/main/resources/reference.conf +++ b/core/src/main/resources/reference.conf @@ -9,7 +9,7 @@ akka.kafka.producer { discovery-method = akka.discovery # Set a service name for use with Akka Discovery - # https://doc.akka.io/docs/alpakka-kafka/current/discovery.html + # https://doc.akka.io/libraries/alpakka-kafka/current/discovery.html service-name = "" # Timeout for getting a reply from the discovery-method lookup @@ -57,7 +57,7 @@ akka.kafka.consumer { discovery-method = akka.discovery # Set a service name for use with Akka Discovery - # https://doc.akka.io/docs/alpakka-kafka/current/discovery.html + # https://doc.akka.io/libraries/alpakka-kafka/current/discovery.html service-name = "" # Timeout for getting a reply from the discovery-method lookup @@ -156,7 +156,7 @@ akka.kafka.consumer { # then causes the Kafka consumer to follow its normal 'auto.offset.reset' behavior. For 'earliest', these settings # allow the client to detect and attempt to recover from this issue. For 'none' and 'latest', these settings will # only add overhead. See - # https://doc.akka.io/docs/alpakka-kafka/current/errorhandling.html#unexpected-consumer-offset-reset + # https://doc.akka.io/libraries/alpakka-kafka/current/errorhandling.html#unexpected-consumer-offset-reset # for more information offset-reset-protection { # turns on reset protection diff --git a/core/src/main/scala/akka/kafka/internal/ConsumerResetProtection.scala b/core/src/main/scala/akka/kafka/internal/ConsumerResetProtection.scala index acab5a73e..7d60b0c1d 100644 --- a/core/src/main/scala/akka/kafka/internal/ConsumerResetProtection.scala +++ b/core/src/main/scala/akka/kafka/internal/ConsumerResetProtection.scala @@ -110,14 +110,14 @@ object ConsumerResetProtection { log.warning( s"Your last commit request $previouslyCommitted is more than the configured threshold from the last" + s"committed offset ($committed) for $tp. See " + - "https://doc.akka.io/docs/alpakka-kafka/current/errorhandling.html#setting-offset-threshold-appropriately for more info." + "https://doc.akka.io/libraries/alpakka-kafka/current/errorhandling.html#setting-offset-threshold-appropriately for more info." ) } log.warning( s"Dropping offsets for partition $tp - received an offset which is less than allowed $threshold " + s"from the last requested offset (threshold: $threshold). Seeking to the latest known safe (committed " + s"or assigned) offset: $committed. See " + - "https://doc.akka.io/docs/alpakka-kafka/current/errorhandling.html#unexpected-consumer-offset-reset" + + "https://doc.akka.io/libraries/alpakka-kafka/current/errorhandling.html#unexpected-consumer-offset-reset" + "for more information." ) consumer ! Seek(Map(tp -> committed.offset())) diff --git a/core/src/main/scala/akka/kafka/javadsl/Consumer.scala b/core/src/main/scala/akka/kafka/javadsl/Consumer.scala index f27dfcc3b..7300299fc 100644 --- a/core/src/main/scala/akka/kafka/javadsl/Consumer.scala +++ b/core/src/main/scala/akka/kafka/javadsl/Consumer.scala @@ -173,7 +173,7 @@ object Consumer { * This is useful when "at-least once delivery" is desired, as each message will likely be * delivered one time but in failure cases could be duplicated. * - * It is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * It is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Producer.flowWithContext]]. */ @ApiMayChange @@ -198,7 +198,7 @@ object Consumer { * This is useful when "at-least once delivery" is desired, as each message will likely be * delivered one time but in failure cases could be duplicated. * - * It is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * It is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Producer.flowWithContext]]. * * This variant makes it possible to add additional metadata (in the form of a string) diff --git a/core/src/main/scala/akka/kafka/javadsl/Producer.scala b/core/src/main/scala/akka/kafka/javadsl/Producer.scala index 706ac055f..df9e669af 100644 --- a/core/src/main/scala/akka/kafka/javadsl/Producer.scala +++ b/core/src/main/scala/akka/kafka/javadsl/Producer.scala @@ -217,7 +217,7 @@ object Producer { * * - [[akka.kafka.ProducerMessage.PassThroughMessage PassThroughMessage]] does not publish anything, and continues in the stream as [[akka.kafka.ProducerMessage.PassThroughResult PassThroughResult]] * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html). + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html). * * @tparam C the flow context type */ @@ -286,7 +286,7 @@ object Producer { * * - [[akka.kafka.ProducerMessage.PassThroughMessage PassThroughMessage]] does not publish anything, and continues in the stream as [[akka.kafka.ProducerMessage.PassThroughResult PassThroughResult]] * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html). + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html). * * Supports sharing a Kafka Producer instance. * diff --git a/core/src/main/scala/akka/kafka/javadsl/Transactional.scala b/core/src/main/scala/akka/kafka/javadsl/Transactional.scala index 1996e0437..b400920d0 100644 --- a/core/src/main/scala/akka/kafka/javadsl/Transactional.scala +++ b/core/src/main/scala/akka/kafka/javadsl/Transactional.scala @@ -39,7 +39,7 @@ object Transactional { /** * API MAY CHANGE * - * This source is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * This source is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Transactional.flowWithOffsetContext]]. */ @ApiMayChange @@ -151,7 +151,7 @@ object Transactional { * carries [[ConsumerMessage.PartitionOffset]] as context. The flow requires a unique `transactional.id` across all app * instances. The flow will override producer properties to enable Kafka exactly-once transactional support. * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Transactional.sourceWithOffsetContext]]. */ @ApiMayChange diff --git a/core/src/main/scala/akka/kafka/scaladsl/Consumer.scala b/core/src/main/scala/akka/kafka/scaladsl/Consumer.scala index 157c626fd..ab36015ea 100644 --- a/core/src/main/scala/akka/kafka/scaladsl/Consumer.scala +++ b/core/src/main/scala/akka/kafka/scaladsl/Consumer.scala @@ -26,7 +26,7 @@ object Consumer { /** * Materialized value of the consumer `Source`. * - * See [[https://doc.akka.io/docs/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] + * See [[https://doc.akka.io/libraries/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] */ trait Control { @@ -36,7 +36,7 @@ object Consumer { * already enqueued messages. It does not unsubscribe from any topics/partitions * as that could trigger a consumer group rebalance. * - * See [[https://doc.akka.io/docs/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] + * See [[https://doc.akka.io/libraries/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] * * Call [[#shutdown]] to close consumer. */ @@ -49,7 +49,7 @@ object Consumer { * from enqueued messages can be handled. * The actor will wait for acknowledgements of the already sent offset commits from the Kafka broker before shutting down. * - * See [[https://doc.akka.io/docs/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] + * See [[https://doc.akka.io/libraries/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] */ def shutdown(): Future[Done] @@ -93,7 +93,7 @@ object Consumer { * one, so that the stream can be stopped in a controlled way without losing * commits. * - * See [[https://doc.akka.io/docs/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] + * See [[https://doc.akka.io/libraries/alpakka-kafka/current/consumer.html#controlled-shutdown Controlled shutdown]] */ final class DrainingControl[T] private (control: Control, val streamCompletion: Future[T]) extends Control { @@ -198,7 +198,7 @@ object Consumer { * This is useful when "at-least once delivery" is desired, as each message will likely be * delivered one time but in failure cases could be duplicated. * - * It is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html), + * It is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html), * [[Producer.flowWithContext]] and/or [[Committer.sinkWithOffsetContext]]. */ @ApiMayChange @@ -219,7 +219,7 @@ object Consumer { * This is useful when "at-least once delivery" is desired, as each message will likely be * delivered one time but in failure cases could be duplicated. * - * It is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html), + * It is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html), * [[Producer.flowWithContext]] and/or [[Committer.sinkWithOffsetContext]]. * * This variant makes it possible to add additional metadata (in the form of a string) diff --git a/core/src/main/scala/akka/kafka/scaladsl/DiscoverySupport.scala b/core/src/main/scala/akka/kafka/scaladsl/DiscoverySupport.scala index 2525aab06..bb1903b18 100644 --- a/core/src/main/scala/akka/kafka/scaladsl/DiscoverySupport.scala +++ b/core/src/main/scala/akka/kafka/scaladsl/DiscoverySupport.scala @@ -120,7 +120,7 @@ object DiscoverySupport { system.dynamicAccess.getClassFor[AnyRef]("akka.discovery.Discovery$") match { case Failure(_: ClassNotFoundException | _: NoClassDefFoundError) => throw new IllegalStateException( - s"Akka Discovery is being used but the `akka-discovery` library is not on the classpath, it must be added explicitly. See https://doc.akka.io/docs/alpakka-kafka/current/discovery.html" + s"Akka Discovery is being used but the `akka-discovery` library is not on the classpath, it must be added explicitly. See https://doc.akka.io/libraries/alpakka-kafka/current/discovery.html" ) case _ => } diff --git a/core/src/main/scala/akka/kafka/scaladsl/Producer.scala b/core/src/main/scala/akka/kafka/scaladsl/Producer.scala index eed8bc42f..df94fe36a 100644 --- a/core/src/main/scala/akka/kafka/scaladsl/Producer.scala +++ b/core/src/main/scala/akka/kafka/scaladsl/Producer.scala @@ -215,7 +215,7 @@ object Producer { * * - [[akka.kafka.ProducerMessage.PassThroughMessage PassThroughMessage]] does not publish anything, and continues in the stream as [[akka.kafka.ProducerMessage.PassThroughResult PassThroughResult]] * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html). + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html). * * @tparam C the flow context type */ @@ -286,7 +286,7 @@ object Producer { * * - [[akka.kafka.ProducerMessage.PassThroughMessage PassThroughMessage]] does not publish anything, and continues in the stream as [[akka.kafka.ProducerMessage.PassThroughResult PassThroughResult]] * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html). + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html). * * Supports sharing a Kafka Producer instance. * diff --git a/core/src/main/scala/akka/kafka/scaladsl/Transactional.scala b/core/src/main/scala/akka/kafka/scaladsl/Transactional.scala index 447faca22..8a112f3d1 100644 --- a/core/src/main/scala/akka/kafka/scaladsl/Transactional.scala +++ b/core/src/main/scala/akka/kafka/scaladsl/Transactional.scala @@ -42,7 +42,7 @@ object Transactional { /** * API MAY CHANGE * - * This source is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * This source is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Transactional.flowWithOffsetContext]]. */ @ApiMayChange @@ -152,7 +152,7 @@ object Transactional { * carries [[ConsumerMessage.PartitionOffset]] as context. The flow requires a unique `transactional.id` across all app * instances. The flow will override producer properties to enable Kafka exactly-once transactional support. * - * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/docs/akka/current/stream/operators/Flow/asFlowWithContext.html) + * This flow is intended to be used with Akka's [flow with context](https://doc.akka.io/libraries/akka-core/current/stream/operators/Flow/asFlowWithContext.html) * and [[Transactional.sourceWithOffsetContext]]. */ @ApiMayChange diff --git a/docs/release-train-issue-template.md b/docs/release-train-issue-template.md index bb741ebd6..335706fbb 100644 --- a/docs/release-train-issue-template.md +++ b/docs/release-train-issue-template.md @@ -30,7 +30,7 @@ Key links: ### Check availability - [ ] Check [API](https://doc.akka.io/api/alpakka-kafka/$VERSION$/) documentation -- [ ] Check [reference](https://doc.akka.io/docs/alpakka-kafka/$VERSION$/) documentation. Check that the reference docs were deployed and show a version warning (see section below on how to fix the version warning). +- [ ] Check [reference](https://doc.akka.io/libraries/alpakka-kafka/$VERSION$/) documentation. Check that the reference docs were deployed and show a version warning (see section below on how to fix the version warning). - [ ] Check the release on https://repo.akka.io/maven/com/typesafe/akka/akka-stream-kafka_2.13/$VERSION$/akka-stream-kafka_2.13-$VERSION$.pom ### When everything is on https://repo.akka.io/maven @@ -40,7 +40,7 @@ Key links: ``` cd ~/www git status - git add docs/alpakka-kafka/current docs/alpakka-kafka/$VERSION$ + git add libraries/alpakka-kafka/current libraries/alpakka-kafka/$VERSION$ git add api/alpakka-kafka/current api/alpakka-kafka/$VERSION$ git commit -m "Alpakka Kafka $VERSION$" ``` @@ -60,7 +60,7 @@ For minor or major releases: ### Afterwards - [ ] Update Alpakka Kafka dependency in Alpakka main repository -- [ ] Update [akka-dependencies bom](https://github.com/akka/akka-dependencies) and version for [Akka module versions](https://doc.akka.io/docs/akka-dependencies/current/) in [akka-dependencies repo](https://github.com/akka/akka-dependencies) -- [ ] Update [Akka Guide samples](https://github.com/akka/akka-platform-guide) +- [ ] Update akka-dependencies bom and version for [Akka module versions](https://doc.akka.io/libraries/akka-dependencies/current/) in [akka-dependencies repo](https://github.com/akka/akka-dependencies) +- [ ] Update [Akka Guide samples](https://github.com/lightbend/akka-guide) - Close this issue diff --git a/docs/src/main/paradox/.htaccess b/docs/src/main/paradox/.htaccess index a9f7771df..3a847fe6a 100644 --- a/docs/src/main/paradox/.htaccess +++ b/docs/src/main/paradox/.htaccess @@ -1,2 +1,2 @@ # No patch version redirect -RedirectMatch 301 ^/docs/alpakka-kafka/([^/]+)/release-notes/1.0.html https://doc.akka.io/docs/alpakka-kafka/$1/release-notes/1.0.x.html +RedirectMatch 301 ^/docs/alpakka-kafka/([^/]+)/release-notes/1.0.html https://doc.akka.io/libraries/alpakka-kafka/$1/release-notes/1.0.x.html diff --git a/docs/src/main/paradox/home.md b/docs/src/main/paradox/home.md index 6687a1f8f..e2d124f6c 100644 --- a/docs/src/main/paradox/home.md +++ b/docs/src/main/paradox/home.md @@ -1,6 +1,6 @@ # Overview -The [Alpakka project](https://doc.akka.io/docs/alpakka/current/) is an open source initiative to implement stream-aware and reactive integration pipelines for Java and Scala. It is built on top of @extref[Akka Streams](akka:stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK @extref[java.util.concurrent.Flow](java-docs:docs/api/java.base/java/util/concurrent/Flow.html)-compliant implementation and therefore @extref[fully interoperable](akka:general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. +The [Alpakka project](https://doc.akka.io/libraries/alpakka/current/) is an open source initiative to implement stream-aware and reactive integration pipelines for Java and Scala. It is built on top of @extref[Akka Streams](akka:stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK @extref[java.util.concurrent.Flow](java-docs:docs/api/java.base/java/util/concurrent/Flow.html)-compliant implementation and therefore @extref[fully interoperable](akka:general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. This **Alpakka Kafka connector** lets you connect [Apache Kafka](https://kafka.apache.org/) to Akka Streams. It was formerly known as **Akka Streams Kafka** and even **Reactive Kafka**. @@ -18,7 +18,7 @@ See all releases in [GitHub releases](https://github.com/akka/alpakka-kafka/rele | 3.5.1 | 3.3, 2.13 | 2.9.0+ | [release 5.0.0](https://github.com/akka/alpakka-kafka/releases/tag/v5.0.0) | 3.3.1 | 2.13, 2.12 | 2.7.0+ | [release 4.0.0](https://github.com/akka/alpakka-kafka/releases/tag/v4.0.0) | 3.0.1 | 2.13 | 2.6.18+ | [release 3.0.1](https://github.com/akka/alpakka-kafka/releases/tag/v3.0.0) -| [3.0.0](https://blogs.apache.org/kafka/entry/what-s-new-in-apache6) | 2.13 | 2.6.18+ | [release 3.0.0 RC1](https://github.com/akka/alpakka-kafka/releases/tag/v3.0.0-RC1) +| [3.0.0](https://blogsarchive.apache.org/kafka/entry/what-s-new-in-apache6) | 2.13 | 2.6.18+ | [release 3.0.0 RC1](https://github.com/akka/alpakka-kafka/releases/tag/v3.0.0-RC1) | [2.7.0](https://archive.apache.org/dist/kafka/2.7.0/RELEASE_NOTES.html) | 2.13, 2.12 | 2.6.14+ | @ref:[release 2.1.0](release-notes/2.1.x.md) | [2.4.1](https://archive.apache.org/dist/kafka/2.4.1/RELEASE_NOTES.html) | 2.13, 2.12, 2.11 | 2.5.31+, 2.6.6+ | @ref:[release 2.0.5](release-notes/2.0.x.md) | [2.4.1](https://archive.apache.org/dist/kafka/2.4.1/RELEASE_NOTES.html) | 2.13, 2.12, 2.11 | 2.5.30+, 2.6.6+ | @ref:[release 2.0.4](release-notes/2.0.x.md) @@ -92,7 +92,7 @@ that are producing or consuming messages to/from Kafka. A few self-contained examples using Alpakka are available as [Alpakka Samples](https://akka.io/alpakka-samples/). -To read and see how others use Alpakka see the [Alpakka documentation's Webinars, Presentations and Articles](https://doc.akka.io/docs/alpakka/current/other-docs/webinars-presentations-articles.html) listing. +To read and see how others use Alpakka see the [Alpakka documentation's Webinars, Presentations and Articles](https://doc.akka.io/libraries/alpakka/current/other-docs/webinars-presentations-articles.html) listing. ## Contributing diff --git a/docs/src/main/paradox/index.md b/docs/src/main/paradox/index.md index 69f5ca87c..85084e0a8 100644 --- a/docs/src/main/paradox/index.md +++ b/docs/src/main/paradox/index.md @@ -1,6 +1,6 @@ # Alpakka Kafka Documentation -The [Alpakka project](https://doc.akka.io/docs/alpakka/current/) is an open source initiative to implement stream-aware and reactive integration pipelines for Java and Scala. It is built on top of @extref[Akka Streams](akka:stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK 9+ @extref[java.util.concurrent.Flow](java-docs:docs/api/java.base/java/util/concurrent/Flow.html)-compliant implementation and therefore @extref[fully interoperable](akka:general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. +The [Alpakka project](https://doc.akka.io/libraries/alpakka/current/) is an open source initiative to implement stream-aware and reactive integration pipelines for Java and Scala. It is built on top of @extref[Akka Streams](akka:stream/index.html), and has been designed from the ground up to understand streaming natively and provide a DSL for reactive and stream-oriented programming, with built-in support for backpressure. Akka Streams is a [Reactive Streams](https://www.reactive-streams.org/) and JDK 9+ @extref[java.util.concurrent.Flow](java-docs:docs/api/java.base/java/util/concurrent/Flow.html)-compliant implementation and therefore @extref[fully interoperable](akka:general/stream/stream-design.html#interoperation-with-other-reactive-streams-implementations) with other implementations. This **Alpakka Kafka connector** lets you connect [Apache Kafka](https://kafka.apache.org/) to Akka Streams. It was formerly known as **Akka Streams Kafka** and even **Reactive Kafka**. diff --git a/docs/src/main/paradox/release-notes/1.0.x.md b/docs/src/main/paradox/release-notes/1.0.x.md index 54be25e5d..d29b90826 100644 --- a/docs/src/main/paradox/release-notes/1.0.x.md +++ b/docs/src/main/paradox/release-notes/1.0.x.md @@ -16,7 +16,7 @@ Released: 2019-07-25 * `FlowWithContext` support [#780](https://github.com/akka/alpakka-kafka/pull/780) * Added connection to broker status checker [#674](https://github.com/akka/alpakka-kafka/issues/674) -This release is compiled and tested against Akka [2.5](https://doc.akka.io/docs/akka/2.5/), [2.6](https://akka.io/blog/news/2019/04/12/akka-2.6-roadmap) and Scala 2.11, 2.12, 2.13.0 on Adopt OpenJDK 8 and 11. +This release is compiled and tested against Akka [2.5](https://doc.akka.io/libraries/akka-core/2.5/), [2.6](https://akka.io/blog/news/2019/04/12/akka-2.6-roadmap) and Scala 2.11, 2.12, 2.13.0 on Adopt OpenJDK 8 and 11. This release was made possible by: @@ -40,7 +40,7 @@ Alpakka Kafka 1.0.4 is released for Scala 2.13, 2.12 and 2.11. The detailed list of changes is found in [the milestone](https://github.com/akka/alpakka-kafka/milestone/32?closed=1). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) and Scala 2.11, 2.12, 2.13.0 on Adopt OpenJDK 8 and 11. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) and Scala 2.11, 2.12, 2.13.0 on Adopt OpenJDK 8 and 11. This release was made possible by: @@ -70,7 +70,7 @@ Szymon Matejczyk, @szymonm](https://github.com/szymonm) The detailed list of changes is found in [the milestone](https://github.com/akka/alpakka-kafka/milestone/31?closed=1). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) and Scala 2.11 and 2.12 on Adopt OpenJDK 1.8. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) and Scala 2.11 and 2.12 on Adopt OpenJDK 1.8. This release was made possible by: @@ -93,11 +93,11 @@ Alpakka Kafka 1.0.2 makes the transactional support more reliable. * Transactions: Add offsets to be committed directly after `producer.send` @github[#752](#752) * Fix transaction offsets for transactional producer @github[#742](#742) by [Szymon Matejczyk, @szymonm](https://github.com/szymonm) * Do not fail commits during a rebalance (avoiding `CommitFailedException`) @github[#755](#755) -* [Testcontainers](https://www.testcontainers.org/) support in the @ref:[Alpakka Kafka testkit](../testing-testcontainers.md#testing-with-a-docker-kafka-cluster) @github[#775](#775) +* [Testcontainers](https://java.testcontainers.org/) support in the @ref:[Alpakka Kafka testkit](../testing-testcontainers.md#testing-with-a-docker-kafka-cluster) @github[#775](#775) The detailed list of changes is found in [the milestone](https://github.com/akka/alpakka-kafka/milestone/30?closed=1). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) and Scala 2.11 and 2.12 on Adopt OpenJDK 1.8. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) and Scala 2.11 and 2.12 on Adopt OpenJDK 1.8. This release was made possible by: @@ -121,7 +121,7 @@ Full [Apache Kafka 2.1.1 release notes](https://archive.apache.org/dist/kafka/2. The upgrade to the Apache Kafka client 2.1.1 is the only change compared to release @ref[1.0](#1-0). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) and Scala 2.11 and 2.12. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) and Scala 2.11 and 2.12. This release was made possible by: @@ -141,9 +141,9 @@ Final release of Alpakka Kafka 1.0! Theses release notes contain the need-to-kno Alpakka Kafka has made a long journey. It started off as **reactive kafka** with a first release in January 2015 when Akka Streams were still experimental and created by [Krzysiek Ciesielski, @kciesielski](https://github.com/kciesielski). The team at [SoftwareMill](https://softwaremill.com/) were the first stewards of the project. -In 2016 the Akka team at [Lightbend](https://www.lightbend.com/) (at that time still called *Typesafe*) took over the responsibility to take care of it and started offering commercial support from version 0.16 (released May 2017). They were assisted by [Alexey Romanchuk, @13h3r](https://github.com/13h3r) during that time. The Akka team has participated significantly in the creation of the [Reactive Streams](https://www.reactive-streams.org/) industry standard and [Akka Streams](https://doc.akka.io/docs/akka/current/stream/index.html) has been one of its first implementations. +In 2016 the Akka team at [Lightbend](https://www.lightbend.com/) (at that time still called *Typesafe*) took over the responsibility to take care of it and started offering commercial support from version 0.16 (released May 2017). They were assisted by [Alexey Romanchuk, @13h3r](https://github.com/13h3r) during that time. The Akka team has participated significantly in the creation of the [Reactive Streams](https://www.reactive-streams.org/) industry standard and [Akka Streams](https://doc.akka.io/libraries/akka-core/current/stream/index.html) has been one of its first implementations. -In May 2018 Lightbend started the Alpakka team to steward the [Alpakka project](https://doc.akka.io/docs/alpakka/current/) and Alpakka Kafka. When the work of improving the tests and the code for Alpakka Kafka kicked off, we dropped the name it had for a while **Akka Streams Kafka**. +In May 2018 Lightbend started the Alpakka team to steward the [Alpakka project](https://doc.akka.io/libraries/alpakka/current/) and Alpakka Kafka. When the work of improving the tests and the code for Alpakka Kafka kicked off, we dropped the name it had for a while **Akka Streams Kafka**. [Apache Kafka](https://kafka.apache.org/) has made an incredible journey during this period. From in-house project at LinkedIn via being open-sourced to becoming the integration backbone of many micro-service architectures. @@ -153,7 +153,7 @@ The need for streaming data has been the key driver for all these endeavours. Alpakka Kafka provides Apache Kafka connectivity for Akka Streams. It supports consuming messages from Kafka into Akka Streams with at-most-once, at-least-once and transactional semantics, and supports producing messages to Kafka. -Once consumed messages are in the Akka Stream, the whole flexibility of all [Akka Stream operators](https://doc.akka.io/docs/akka/current/stream/operators/index.html) becomes available. +Once consumed messages are in the Akka Stream, the whole flexibility of all [Akka Stream operators](https://doc.akka.io/libraries/akka-core/current/stream/operators/index.html) becomes available. Alpakka Kafka achieves back-pressure for consuming by automatically pausing and resuming its Kafka subscriptions. When there is no downstream demand for more data, the consumer will not read any data. Any other communication with the Kafka broker (heartbeats, committing, rebalancing) will still continue. diff --git a/docs/src/main/paradox/release-notes/1.1.x.md b/docs/src/main/paradox/release-notes/1.1.x.md index 0c91dd94c..45523b750 100644 --- a/docs/src/main/paradox/release-notes/1.1.x.md +++ b/docs/src/main/paradox/release-notes/1.1.x.md @@ -34,7 +34,7 @@ Released: 2019-10-02 The detailed list of changes is found in [the 1.1.0-RC2 Github release listing](https://github.com/akka/alpakka-kafka/releases/tag/v1.1.0-RC2). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) for Scala 2.11, 2.12 and 2.13, [Akka 2.6-M8](https://doc.akka.io/docs/akka/2.6/) for Scala 2.12, and 2.13 and both Akka versions on Adopt OpenJDK 1.8 and 11. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) for Scala 2.11, 2.12 and 2.13, [Akka 2.6-M8](https://doc.akka.io/libraries/akka-core/2.6/) for Scala 2.12, and 2.13 and both Akka versions on Adopt OpenJDK 1.8 and 11. **If we do not get any reports that speak against it, we'll release Alpakka Kafka 1.1.0 in a few weeks.** @@ -68,7 +68,7 @@ This goes along with a few improvements to the documentation. The detailed list of changes is found in [the 1.1.0-RC1 Github release listing](https://github.com/akka/alpakka-kafka/releases/tag/v1.1.0-RC1). -This release is compiled and tested against [Akka 2.5](https://doc.akka.io/docs/akka/2.5/) for Scala 2.11, 2.12 and 2.13, [Akka 2.6-M6](https://doc.akka.io/docs/akka/2.6/) for Scala 2.12, and 2.13 and both Akka versions on Adopt OpenJDK 1.8 and 11. +This release is compiled and tested against [Akka 2.5](https://doc.akka.io/libraries/akka-core/2.5/) for Scala 2.11, 2.12 and 2.13, [Akka 2.6-M6](https://doc.akka.io/libraries/akka-core/2.6/) for Scala 2.12, and 2.13 and both Akka versions on Adopt OpenJDK 1.8 and 11. This release contains contributions by: diff --git a/docs/src/main/paradox/release-notes/2.0.x.md b/docs/src/main/paradox/release-notes/2.0.x.md index 831cacce8..354ad43e7 100644 --- a/docs/src/main/paradox/release-notes/2.0.x.md +++ b/docs/src/main/paradox/release-notes/2.0.x.md @@ -352,7 +352,7 @@ Released: 2019-12-03 The detailed list of changes is found in [the 2.0.0-M2 Github release listing](https://github.com/akka/alpakka-kafka/releases/tag/v2.0.0-M2). -This release is compiled and tested against Akka [2.5.23](https://doc.akka.io/docs/akka/2.5/), [2.6.0](https://doc.akka.io/docs/akka/2.6/) and Scala 2.11, 2.12, 2.13 on Adopt OpenJDK 8 and 11. +This release is compiled and tested against Akka [2.5.23](https://doc.akka.io/libraries/akka-core/2.5/), [2.6.0](https://doc.akka.io/libraries/akka-core/2.6/) and Scala 2.11, 2.12, 2.13 on Adopt OpenJDK 8 and 11. This release contains contributions by: diff --git a/docs/src/main/paradox/serialization.md b/docs/src/main/paradox/serialization.md index 875a5f4b6..90dd912a9 100644 --- a/docs/src/main/paradox/serialization.md +++ b/docs/src/main/paradox/serialization.md @@ -5,7 +5,7 @@ The general recommendation for de-/serialization of messages is to use byte arra ## Protocol buffers -[Protocol Buffers](https://developers.google.com/protocol-buffers) offer a language-neutral, platform-neutral, extensible mechanism for serializing structured data and allow consumers and producers to rely on the message format. +[Protocol Buffers](https://protobuf.dev) offer a language-neutral, platform-neutral, extensible mechanism for serializing structured data and allow consumers and producers to rely on the message format. The easiest way to use Protocol Buffers with Alpakka Kafka is to serialize and deserialize the Kafka message payload as a byte array and call the Protocol Buffers serialization and deserialization in a regular `map` operator. To serialize the Protobuf-defined type `Order` into a byte array use the `.toByteArray()` method which gets generated by the Protobuf compiler. diff --git a/docs/src/main/paradox/snapshots.md b/docs/src/main/paradox/snapshots.md index c6368f23e..0635d898f 100644 --- a/docs/src/main/paradox/snapshots.md +++ b/docs/src/main/paradox/snapshots.md @@ -41,5 +41,5 @@ Gradle ## Documentation -The [snapshot documentation](https://doc.akka.io/docs/alpakka-kafka/snapshot/) is updated with every snapshot build. +The [snapshot documentation](https://doc.akka.io/libraries/alpakka-kafka/snapshot/) is updated with every snapshot build. diff --git a/docs/src/main/paradox/testing-testcontainers.md b/docs/src/main/paradox/testing-testcontainers.md index 5b65111c8..209d433e1 100644 --- a/docs/src/main/paradox/testing-testcontainers.md +++ b/docs/src/main/paradox/testing-testcontainers.md @@ -3,7 +3,7 @@ project.description: Alpakka Kafka provides Testcontainers support for running a --- # Testing with a Docker Kafka cluster -The [Testcontainers](https://www.testcontainers.org/) project contains a nice API to start and stop Apache Kafka in Docker containers. +The [Testcontainers](https://java.testcontainers.org/) project contains a nice API to start and stop Apache Kafka in Docker containers. This becomes very relevant when your application code uses a Scala version which Apache Kafka doesn't support so that *EmbeddedKafka* can't be used. Testcontainers also allow you to create a complete Kafka cluster (using Docker containers) to simulate production use cases. @@ -53,7 +53,7 @@ You can retrieve the Schema Registry URL in your test configuration by calling ` ## Testing with a Docker Kafka cluster from Java code -The Alpakka Kafka testkit contains helper classes to start Kafka via Testcontainers. Alternatively, you may use just Testcontainers, as it is designed to be used with JUnit and you can follow [their documentation](https://www.testcontainers.org/modules/kafka/) to start and stop Kafka. To start a single instance for many tests see [Singleton containers](https://www.testcontainers.org/test_framework_integration/manual_lifecycle_control/). +The Alpakka Kafka testkit contains helper classes to start Kafka via Testcontainers. Alternatively, you may use just Testcontainers, as it is designed to be used with JUnit and you can follow [their documentation](https://java.testcontainers.org/modules/kafka/) to start and stop Kafka. To start a single instance for many tests see [Singleton containers](https://java.testcontainers.org/test_framework_integration/manual_lifecycle_control/). The Testcontainers dependency must be added to your project explicitly. diff --git a/project/project-info.conf b/project/project-info.conf index 9ca849f73..b5bd5d336 100644 --- a/project/project-info.conf +++ b/project/project-info.conf @@ -12,7 +12,7 @@ project-info { text: "Github issues" } release-notes: { - url: "https://doc.akka.io/docs/alpakka-kafka/current/release-notes/index.html" + url: "https://doc.akka.io/libraries/alpakka-kafka/current/release-notes/index.html" text: "In the documentation" new-tab: false } diff --git a/scripts/link-validator.conf b/scripts/link-validator.conf index f7eb829ee..4d9c53aab 100644 --- a/scripts/link-validator.conf +++ b/scripts/link-validator.conf @@ -2,13 +2,13 @@ site-link-validator { root-dir = "./docs/target/site/" # relative to `root-dir` - start-file = "docs/alpakka-kafka/snapshot/index.html" + start-file = "libraries/alpakka-kafka/snapshot/index.html" # Resolves URLs with the given prefix as local files instead link-mappings = [ { - prefix = "https://doc.akka.io/docs/alpakka-kafka/snapshot/" - replace = "/docs/alpakka-kafka/snapshot/" + prefix = "https://doc.akka.io/libraries/alpakka-kafka/snapshot/" + replace = "/libraries/alpakka-kafka/snapshot/" } { prefix = "https://doc.akka.io/api/alpakka-kafka/snapshot/" @@ -18,6 +18,8 @@ site-link-validator { ignore-missing-local-files-regex = "" + ignore-files = [] + ignore-prefixes = [ # runtime is NOT part of the published Scaladoc "https://www.scala-lang.org/api/2.13.14/scala/runtime/AbstractFunction2.html" diff --git a/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaJunit4Test.java b/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaJunit4Test.java index 76ea14dbf..0d3d6748f 100644 --- a/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaJunit4Test.java +++ b/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaJunit4Test.java @@ -14,7 +14,7 @@ import org.junit.Before; /** - * JUnit 4 base class using [[https://www.testcontainers.org/ Testcontainers]] to start a Kafka + * JUnit 4 base class using [[https://java.testcontainers.org/ Testcontainers]] to start a Kafka * broker in a Docker container. The Kafka broker will be kept around across multiple test classes, * unless `stopKafka()` is called. * diff --git a/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaTest.java b/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaTest.java index 364fe445f..db712cdc7 100644 --- a/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaTest.java +++ b/testkit/src/main/java/akka/kafka/testkit/javadsl/TestcontainersKafkaTest.java @@ -12,7 +12,7 @@ import akka.stream.Materializer; /** - * JUnit 5 base class using [[https://www.testcontainers.org/ Testcontainers]] to start a Kafka + * JUnit 5 base class using [[https://java.testcontainers.org/ Testcontainers]] to start a Kafka * broker in a Docker container. The Kafka broker will be kept around across multiple test classes, * unless `stopKafka()` is called (eg. from an `@AfterAll`-annotated method. * diff --git a/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaLike.scala b/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaLike.scala index b6033eb5b..c007de112 100644 --- a/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaLike.scala +++ b/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaLike.scala @@ -10,7 +10,7 @@ import akka.kafka.testkit.internal.{AlpakkaKafkaContainer, SchemaRegistryContain import org.testcontainers.containers.GenericContainer /** - * Uses [[https://www.testcontainers.org/ Testcontainers]] to start a Kafka cluster in a Docker container. + * Uses [[https://java.testcontainers.org/ Testcontainers]] to start a Kafka cluster in a Docker container. * This trait will start Kafka only once per test session. To create a Kafka cluster per test class see * [[TestcontainersKafkaPerClassLike]]. * diff --git a/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaPerClassLike.scala b/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaPerClassLike.scala index 5b97f096d..11aa6cdd1 100644 --- a/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaPerClassLike.scala +++ b/testkit/src/main/scala/akka/kafka/testkit/scaladsl/TestcontainersKafkaPerClassLike.scala @@ -8,7 +8,7 @@ package akka.kafka.testkit.scaladsl import akka.kafka.testkit.internal.TestcontainersKafka /** - * Uses [[https://www.testcontainers.org/ Testcontainers]] to start a Kafka broker in a Docker container once per class. + * Uses [[https://java.testcontainers.org/ Testcontainers]] to start a Kafka broker in a Docker container once per class. * The Testcontainers dependency has to be added explicitly. */ trait TestcontainersKafkaPerClassLike extends TestcontainersKafka.Spec { diff --git a/tests/src/test/scala/akka/kafka/scaladsl/IntegrationSpec.scala b/tests/src/test/scala/akka/kafka/scaladsl/IntegrationSpec.scala index edd9dd67b..a46f075c6 100644 --- a/tests/src/test/scala/akka/kafka/scaladsl/IntegrationSpec.scala +++ b/tests/src/test/scala/akka/kafka/scaladsl/IntegrationSpec.scala @@ -238,7 +238,7 @@ class IntegrationSpec extends SpecBase with TestcontainersKafkaLike with Inside Await.result(produce(topic, 1 to 100), remainingOrDefault) // Create ConsumerActor manually - // https://doc.akka.io/docs/akka-stream-kafka/0.20/consumer.html#sharing-kafkaconsumer + // https://doc.akka.io/libraries/akka-stream-kafka/0.20/consumer.html#sharing-kafkaconsumer val consumer = system.actorOf(KafkaConsumerActor.props(consumerDefaults.withGroupId(group))) // Timeout for metadata fetching requests diff --git a/tests/src/test/scala/akka/kafka/tests/CapturingAppender.scala b/tests/src/test/scala/akka/kafka/tests/CapturingAppender.scala index 9b5fb3b59..f558902d4 100644 --- a/tests/src/test/scala/akka/kafka/tests/CapturingAppender.scala +++ b/tests/src/test/scala/akka/kafka/tests/CapturingAppender.scala @@ -10,7 +10,7 @@ import ch.qos.logback.classic.spi.ILoggingEvent import ch.qos.logback.core.AppenderBase /** - * See https://doc.akka.io/docs/akka/current/typed/testing-async.html#silence-logging-output-from-tests + * See https://doc.akka.io/libraries/akka-core/current/typed/testing-async.html#silence-logging-output-from-tests * * INTERNAL API */ @@ -35,7 +35,7 @@ import ch.qos.logback.core.AppenderBase } /** - * See https://doc.akka.io/docs/akka/current/typed/testing-async.html#silence-logging-output-from-tests + * See https://doc.akka.io/libraries/akka-core/current/typed/testing-async.html#silence-logging-output-from-tests * * INTERNAL API * diff --git a/tests/src/test/scala/akka/kafka/tests/LogbackUtil.scala b/tests/src/test/scala/akka/kafka/tests/LogbackUtil.scala index 18a8a218d..022c8ef82 100644 --- a/tests/src/test/scala/akka/kafka/tests/LogbackUtil.scala +++ b/tests/src/test/scala/akka/kafka/tests/LogbackUtil.scala @@ -10,7 +10,7 @@ import ch.qos.logback.classic.Level import org.slf4j.LoggerFactory /** - * See https://doc.akka.io/docs/akka/current/typed/testing-async.html#silence-logging-output-from-tests + * See https://doc.akka.io/libraries/akka-core/current/typed/testing-async.html#silence-logging-output-from-tests * * INTERNAL API */ diff --git a/tests/src/test/scala/akka/kafka/tests/javadsl/LogCapturingJunit4.scala b/tests/src/test/scala/akka/kafka/tests/javadsl/LogCapturingJunit4.scala index 29b886e69..c29428258 100644 --- a/tests/src/test/scala/akka/kafka/tests/javadsl/LogCapturingJunit4.scala +++ b/tests/src/test/scala/akka/kafka/tests/javadsl/LogCapturingJunit4.scala @@ -14,7 +14,7 @@ import org.junit.runners.model.Statement import org.slf4j.LoggerFactory /** - * See https://doc.akka.io/docs/akka/current/typed/testing-async.html#silence-logging-output-from-tests + * See https://doc.akka.io/libraries/akka-core/current/typed/testing-async.html#silence-logging-output-from-tests * * JUnit `TestRule` to make log lines appear only when the test failed. * diff --git a/tests/src/test/scala/akka/kafka/tests/scaladsl/LogCapturing.scala b/tests/src/test/scala/akka/kafka/tests/scaladsl/LogCapturing.scala index aad991584..c45f83db5 100644 --- a/tests/src/test/scala/akka/kafka/tests/scaladsl/LogCapturing.scala +++ b/tests/src/test/scala/akka/kafka/tests/scaladsl/LogCapturing.scala @@ -15,7 +15,7 @@ import org.scalatest.TestSuite import org.slf4j.LoggerFactory /** - * See https://doc.akka.io/docs/akka/current/typed/testing-async.html#silence-logging-output-from-tests + * See https://doc.akka.io/libraries/akka-core/current/typed/testing-async.html#silence-logging-output-from-tests * * Mixin this trait to a ScalaTest test to make log lines appear only when the test failed. *