Golang: Go Routines and WaitGroups

mr_destructive

Meet Rajesh Gor

Posted on July 31, 2023

Golang: Go Routines and WaitGroups

Introduction

One of the key features that set Go apart from many other languages is its native support for Goroutines - lightweight concurrent functions that can run concurrently and efficiently manage concurrency tasks. Goroutines are an essential aspect of Go's concurrency model, enabling developers to build highly concurrent and performant applications effortlessly.

In this blog post, we will dive into the world of Goroutines and explore how they work, why they are essential for concurrent programming in Go, and how they can greatly improve the responsiveness and efficiency of your applications. This post will cover go routines and a primer on wait groups, in the next article we will be looking deeply into channels where all these three things can be demonstrated and understood in a better and more useful way.

Concurrency and Parallelism

This two concepts are quite crucial before diving into the fundamentals of concurreny.

Concurrency is about dealing with multiple things at once. Concurrent programs can have several processes or threads running simultaneously on a single CPU core by rapidly switching between them (context switching). The threads are interleaved, not necessarily executing at literally the same time. The CPU can switch between these tasks to give the appearance of simultaneous progress.

Parallelism is about doing multiple things at literally the same time. Parallel programs can execute multiple computations simultaneously on separate CPU cores. The threads actually execute in parallel.

What is a Go Routine

A go routine is a simple lightweight thread managed by the Go runtime. In the simplest of terms, a go routine can be defined as:

Go routine is a way to perform multiple tasks within a program, allowing different parts of the program to work simultaneously and make the most out of the resources.

Also it can be stated as:

A goroutine in Golang is a lightweight, independently executing function that runs concurrently with other goroutines within the same address space. In other words, it is a concurrent unit of execution.

Focussing on the word same address space that will be really critical in the later sections of this article.

Features of Go Routines

Go routines form a key part of Go's concurrency model. Here are some of the key features of go routines:

  • Lightweight Thread:
    A Go routine is often referred to as a lightweight thread.

  • Independent Execution:
    Go routines run independently of each other, enabling concurrent execution.

  • Managed by Go:
    Go routines are managed by the Go runtime, making them easy to use.

  • Low Overhead:
    Go routines have low memory overhead, allowing us to create thousands of them efficiently.

  • Communication:
    Go routines can communicate and synchronize data through channels.

  • Asynchronous:
    Go routines can execute asynchronously, allowing other parts of the program to continue running.

  • Scalability:
    Go routines are the foundation of scalable concurrent programming in Go.

Unlike threads in other languages, goroutines are cheap and you can easily create thousands or even millions of them in a program.

Example of Go Routines

Creating a go routine is not hard, just add the keyword go before the function call and the go runtime will create a new go routine inside the main function or wherever is the context from. Remember, the main function is also a go routine.

package main

import (
    "fmt"
    "time"
)

func process() {
    fmt.Println("Hello World!")
    time.Sleep(time.Second)
}

func main() {
    start := time.Now()
    go process()
    go process()
    end := time.Now()
    duration := end.Sub(start)
    fmt.Println(duration)
}
Enter fullscreen mode Exit fullscreen mode
scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

Hello World!
15.607ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

9.889ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

8.834ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

9.158ยตs

scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

12.54ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

Hello World!
Hello World!
10.19ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ go run main.go 

14.1ยตs


scripts/go-routines on ๎‚  main via ๐Ÿน v1.20 
$ 
Enter fullscreen mode Exit fullscreen mode

Quite unpredictable output right? This is a power of go-routines. It is asynchronous so it will not block the main function. The two function calls to process, are executed independently of the main function scope. The program just calls the process function and captures the output, and sequentially reaches the end of the program(main function), At this point the go routines(threads) inside the main function are stopped abruptly.

Let's break it down if it is not clear yet.

  • The main function starts.
  • go process() creates a Go routine and starts its execution.
  • Meanwhile, another call to go process() creates a separate Go routine and starts its execution as well.
  • Meanwhile, calculate the time difference between the main function start and end.
  • Meanwhile main function ends.

So, in summary, the main function is only able to capture the output of the duration as it is synchronous and if the process has been executed, it prints the Hello World! message. Hence we have different outputs because the uncontrolled concurrency, lack of coordination, and OS scheduling across the program runs are different.

Wait Groups

In simple terms, WaitGroup is used to synchronize multiple goroutines and to wait for them to finish executing. This allows the go routines to be completed before the completion of the main function, hence it blocks the main function from leaving/exiting the scope.

A WaitGroup is a synchronization primitive that allows a goroutine to wait for a collection of other goroutines to finish executing.

  • A WaitGroup is initialized with a counter representing the number of goroutines to wait for.
  • The Add() method increments the counter by the given value. This is called by each goroutine to indicate it is running.
  • The main goroutine calls Add() to set the initial count, then launches worker goroutines.
  • A WaitGroup is typically passed by a pointer to goroutines that need to be waited on.
  • The Done() method decrements the counter by 1. Goroutines call this when finished.
  • Each worker calls Done() when finished, decrementing the counter.
  • The Wait() method blocks until the counter reaches 0, indicating all goroutines have finished.
  • Main calls Wait() to block until Done() brings counter to 0.

This provides a simple way to synchronize multiple goroutines finishing their work with a main thread that needs to wait for them to complete. The counter ensures the main thread knows how many goroutines it is waiting for. Interacting and working with go routines with synchronization using a wait group is quite intuitive and simple to follow. Let's look at a simple example below:

package main

import (
    "fmt"
    "sync"
    "time"
)

func process(pid int, wg *sync.WaitGroup) {
    fmt.Printf("Started process %d\n", pid)
    time.Sleep(1 * time.Second)
    fmt.Printf("Completed process %d\n", pid)
    defer wg.Done()
}

func main() {

    now := time.Now()
    var wg sync.WaitGroup

    for i := 0; i < 10; i++ {
        wg.Add(1)
        go process(i, &wg)
    }
    wg.Wait()
    fmt.Println("All processes completed")
    end := time.Now()
    fmt.Println(end.Sub(now))
}
Enter fullscreen mode Exit fullscreen mode
scripts/go-routines on ๎‚  main [?] via ๐Ÿน v1.20 
$ go run wg.go 
Started process 9
Started process 0
Started process 1
Started process 2
Started process 3
Started process 4
Started process 5
Started process 6
Started process 7
Started process 8
Completed process 8
Completed process 3
Completed process 9
Completed process 0
Completed process 1
Completed process 2
Completed process 5
Completed process 4
Completed process 6
Completed process 7
1.000563257s
All processes completed

scripts/go-routines on ๎‚  main [?] via ๐Ÿน v1.20 
$ 

Enter fullscreen mode Exit fullscreen mode

In the above example, we have used the same function process but with a slight twist, we have added a process id, just an integer to represent the go routine. We print the start and completion of the function in between the sleep for 1 second. We also have a wait group.

  • A WaitGroup is basically a struct type defined in the sync package.
  • The variable wg is a new wait group instance that will be used to synchronize and wait for the completion of groups of go routines.

We create a for loop with 10 iterations, so as to spawn 10 process function calls. We first use the wg.Add(1) which says to the wait group to wait for 1 go routine. The immediate next line is a go routine go process() which takes in a pid just to track which go routine is being executed in the loop.

Inside the process function, we simply say the process with the given pid has started as a print statement, sleeps for a second, and then print the end of the process with pid. The end of the function is marked with defer wg.Done(), this is to indicate that the go routine is completed.

The wg(wait group) has a counter that keeps track of the number of go routines it has to synchronize or wait for them to complete. In the Add function, the internal counter in the WaitGroup is incremented by the delta the integer parsed as a parameter. And the Done function, decrements the internal counter in the WaitGroup which indicates that the go routine is completed.

In the main function, we call the wg.Wait which will block until the counter for the WaitGroup is 0 i.e. all go routines have completed their execution. So, we created 10 go routines which ran concurrently but were synchronized with the help of WaitGroups. The WaitGroup is allowed to block the main function till all the go routines are done executing.

Go Routine with WaitGroup Use Cases

Go routines can be used for creating asynchronous tasks, and also for creating concurrent tasks. By using wait groups, we can create a way to wait for multiple goroutines to complete. By using go routines and wait groups, we can complete n number of tasks in a time frame of 1 task's completion. However, to create concurrent communication between the other processes, we need channels (which we will explore in the next article).

Here is a simple breakdown of what asynchronous and concurrent tasks might refer to:

Asynchronous tasks run independently of the main program flow, allowing the main program to continue executing without waiting for the task's completion. For example, not blocking other tasks in the sequential flow of the main function.

Concurrent tasks run simultaneously and can execute at the same time as other tasks. They make use of multiple threads (goroutines in Go's case) to achieve parallel execution. For example, running multiple tasks parallel will cut the time down for spinning each task after the completion of another task.

Some of the asynchronous tasks that can be done with go routines might include the following:

  • Sending a mail while saving the user to the database.
  • Fetch and process data from multiple websites(web scraping/crawling).
  • High-performance message brokers and queuing systems for inter-process communication.

One practical way of using go routines with wait groups would be to send mail, we are not going to see the actual implementation of the mail-sending stuff. However, we can try to mimic how the setup might be for sending bulk mail. By creating a wait group and a list of mail addresses to send, a go routine can be created with a function that sends those emails.

package main

import (
    "fmt"
    "sync"
)

func sendMail(address string, wg *sync.WaitGroup) {
    fmt.Println("Sending mail to", address)
    defer wg.Done()
    // Actual mail sending, smtp stuff
    // handle errors

    // client, err := smtp.Dial("smtp.example.com:587")
    // errr = client.Mail("sender@example.com")
    // err = client.Rcpt(address)
    // wc, err := client.Data()
    //_, err = wc.Write([]byte("This is the email body."))
    // err = wc.Close()
    // client.Quit()
}

func main() {
    emails := []string{
        "recipient1@example.com",
        "recipient2@example.com",
        "xyz@example.com",
    }
    wg := sync.WaitGroup{}
    wg.Add(len(emails))

    for _, email := range emails {
        go sendMail(email, &wg)
    }
    wg.Wait()
    fmt.Println("All emails queued for sending")
    // Do other stuff
}
Enter fullscreen mode Exit fullscreen mode

In the above example, the emails is a list of email ids to send the mail. We have created a wait group that we can initialize with the total number of go routines probably to be executed. The wg.Add method is parsed with the number of emails to be sent hence that equals the go routines to spawn.

So, we can then in the for loop, iterate over each mail and send the emails with the sendMail function as go routine. The wg.Wait function outside the loop, will make sure the main function is halted before all the go routines complete their execution.

There is one more way to call a function as a go routine without changing its signature since we had to pass wg pointer as the WaitGroup reference to it to acknowledge the completion of the go routine. We could wrap these two operations wiz. calling the function and calling the wg.Done method with an anonymous function.

package main

import (
    "fmt"
    "sync"
)

func sendMail(address string) {
    fmt.Println("Sending mail to", address)
}

func main() {
    emails := []string{
        "recipient1@example.com",
        "recipient2@example.com",
        "xyz@example.com",
    }
    wg := sync.WaitGroup{}
    wg.Add(len(emails))

    for _, email := range emails {
        mail := email
        go func(m string) {
            sendMail(m)
            wg.Done()
        }(mail)
    }
    wg.Wait()
    fmt.Println("All emails queued for sending")
    // Do other stuff
}
Enter fullscreen mode Exit fullscreen mode

This does the exact same thing, but we don't have to change the signature of the function, This keeps the functional logic of the function in its place and handles the concurrency on the go.


for _, email := range emails {
    mail := email
    go func(m string) {
        sendMail(m)
        wg.Done()
    }(mail)
}
Enter fullscreen mode Exit fullscreen mode

If the above bit scares you, don't worry it's too simple.

  • We are iterating over the email slice using a for loop and creating a Go routine for each email address. The loop variable email represents the email address at the current iteration.
  • However, to avoid the loop variable capture issue (where all Go routines would share the same email variable), we create a new variable mail and assign the value of email to it. This step ensures that each Go routine captures its own copy of the email address.
  • We immediately create an anonymous function (a closure) using the go keyword. This anonymous function takes the mail variable as a parameter m and is executed concurrently as a Go routine. Inside the Go routine, we call the sendMail function with the email address m.
  • After the sendMail call has been executed i.e. email has been sent, we call wg.Done() to notify the WaitGroup that the Go routine has completed its work. This allows the WaitGroup to properly synchronize and wait for all Go routines to finish before the program proceeds beyond wg.Wait() in the main function.

This is one way to do it if you want to separate the mail-sending logic from the goroutines/concurrency task. However, this should be handled with care as the variables inside the closure might be shared among all the goroutines instead of having individual variable literals.

To ensure that each goroutine operates on its own copy of the email address, we use the approach of creating a new variable mail and passing it as a parameter to the anonymous function. This way, each goroutine captures its unique email address, avoiding any interference or unintended sharing of data between goroutines.

Mutual Exclusion Locks

In the previous examples, we saw how goroutines and wait groups allow us to run multiple tasks concurrently in Go. However, sometimes these concurrent goroutines need to access shared resources like memory, files, network sockets, etc.

When more than one goroutine tries to access a resource at the same time, it can lead to race conditions and unpredictable behavior. To handle this, we need a way to ensure only one goroutine can access the resource at a time.

This is where mutual exclusion locks come in. A mutual exclusion lock, or mutex, provides a mechanism to lock access to a shared resource. It ensures only one goroutine at a time can acquire the lock, blocking other goroutines until the lock is released.

For example, say we have multiple goroutines trying to append data to the same memory buffer(could be file/database/etc.) concurrently:

package main

import (
    "fmt"
    "os"
    "sync"
)

func WriteToFile(filename string, contents string, buffer *[]byte, wg *sync.WaitGroup) {
    defer wg.Done()
    *buffer = append(*buffer, []byte(contents)...)
    err := os.WriteFile(filename, *buffer, 0644)
    if err != nil {
        fmt.Println(err)
    }
}

func main() {
    var wg sync.WaitGroup
    var sharedBuffer []byte

    wg.Add(2)
    go WriteToFile("data/f1.txt", "Hello ", &sharedBuffer, &wg)
    go WriteToFile("data/f1.txt", "World! ", &sharedBuffer, &wg)
    wg.Wait()

    fmt.Println(string(sharedBuffer))
}
Enter fullscreen mode Exit fullscreen mode
$ go run --race no-mutex.go
==================
WARNING: DATA RACE
Read at 0x00c000012030 by goroutine 8:
  main.WriteToFile()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:11 +0xe7
  main.main.func2()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:24 +0x64

Previous write at 0x00c000012030 by goroutine 7:
  main.WriteToFile()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:11 +0x16a
  main.main.func1()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:23 +0x64

Goroutine 8 (running) created at:
  main.main()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:24 +0x1f6

Goroutine 7 (finished) created at:
  main.main()
      /home/meet/code/100-days-of-golang/scripts/go-routines/no-mutex.go:23 +0x14e
==================
Enter fullscreen mode Exit fullscreen mode

This is because we are trying to access the same memory address at the same time. This is a race condition and can lead to unpredictable behavior. Try removing the --race flag while running, in this little stupid example, it might not be obvious, but in complex and constrained environments, this can get the application in serious trouble.

NOTE: We are using go run --race no-mutex.go to check if there are any race conditions in the program. This is the race detector flag in the run command.

To avoid this race condition, we need to add the mutex locks provided in the sync.Mutex type. There are methods like Lock, Unlock, and TryLock which help in locking access of the resource to a single entity at a given time.

When a goroutine calls Lock() on a mutex, it acquires the lock. If the mutex is already locked by another goroutine, the calling goroutine will be blocked (put to sleep) until the lock becomes available. Once the lock is acquired successfully, the goroutine can proceed with its critical section, which is the part of the code that should not be executed concurrently by multiple goroutines.

When a goroutine calls Unlock() on a mutex, it releases the lock. This allows other waiting goroutines to acquire the lock and execute their critical sections. It's essential to ensure that Unlock() is called after the critical section to release the mutex and avoid deadlocks. The critical section/shared resource should not be accessed after the release of this lock.

package main

import (
    "fmt"
    "os"
    "sync"
)

func WriteToFile(filename string, contents string, buffer *[]byte, wg *sync.WaitGroup, mutex *sync.Mutex) {
    defer wg.Done()
    contentBytes := []byte(contents)

    mutex.Lock()
    *buffer = append(*buffer, contentBytes...)

    f, err := os.OpenFile(filename, os.O_APPEND|os.O_WRONLY, 0644)
    if err != nil {
        fmt.Println(err)
    }
    defer f.Close()
    _, err = f.Write(contentBytes)
    if err != nil {
        fmt.Println(err)
    }
    mutex.Unlock()
}

func main() {
    var wg sync.WaitGroup
    var mut sync.Mutex
    var sharedBuffer []byte

    wg.Add(2)
    go WriteToFile("data/f1.txt", "Hello Gophers!\n", &sharedBuffer, &wg, &mut)
    go WriteToFile("data/f1.txt", "Welcome to Goworld!\n", &sharedBuffer, &wg, &mut)
    wg.Wait()

    fmt.Println(string(sharedBuffer))
}
Enter fullscreen mode Exit fullscreen mode
$ go run --race mutex.go
Welcome to Goworld!
Hello Gophers!

$ go run --race mutex.go
Hello Gophers!
Welcome to Goworld!

$ go run --race mutex.go
Hello Gophers!
Welcome to Goworld!

$ go run --race mutex.go
Welcome to Goworld!
Hello Gophers!
Enter fullscreen mode Exit fullscreen mode

The above example is a preventive measure for race conditions with a mutex Lock on the shared resource used in the go routines.

Let's break down the code step by step. First, we initialize a few variables:

  • wg as sync.WaitGroup:
    The wg is a waitgroup that will be used for the synchronization of the go-routines by blocking the main function to exit

  • mut as sync.Mutex:
    The mut is a structure that internally holds a few integer values indicating either a blocked or unblocked state. The sync.Mutex has two private fields wiz. state and sema, the state holds the mutex state either 0(unlocked) or 1 as locked. The sema field is a uint32 that is used for blocking and signaling, it acts as a semaphore to manage blocking and unblocking goroutines trying to acquire the mutex.
    This will be used to acquire Lock and Unlock on the shared resource while writing the data to the file or appending the data to the resource.

  • sharedBuffer as []byte:
    The sharedBuffer is the actually shared resource that will be used to hold the strings for keeping track of the data written to the file. It will be the resource that would require to lock for mutating its value(appending to the slice) among the go routines.

We add 2 to the wg to indicate to wait for the two go-routines to complete, in the next line we call two go routines as the function WriteToFile. The WriteToFile is a function that takes in quite a few parameters namely, the filename, the content to write, the reference to the sharedBuffer, waitgroup, and the mutex.

Inside the function WriteToFile:

  • We first defer the waitgroups as Done i.e. to call the wg.Done method at the end of the function call.
  • Typecast the contents from string as the []byte.
  • Acquire the mutex.Lock() i.e. to say "The below operations should not be done concurrently, one at a time". We then append the contents to the buffer which is a pointer to the sharedBuffer in the main function, so essentially we are trying to mutate the sharedBuffer in this function.
  • Open the file as O_APPEND and O_WRONLY to indicate that the file should be opened in append/writing mode. (We have observed this type of operation in the Golang: File Write article)
  • Use the Write method to write the slice of bytes(contents) into the file that we opened. We have a defer for the closing of the file.

We obviously check for errors and print them, but it could be a panic or log depending on the type of operation the application is doing.
So that is the all operation we want to do, so we finally open the lock with mutex.Unlock() which will allow the other go routine if any to access the resource and proceed with the operations.

Read Write Mutual Exclusion Lock

The mutual exclusion lock is good if you have a write operation-heavy application. However, if we have read write operations in more or the less in same proportion(read heavy) we don't want to have readers getting blocked when other readers are accessing the resource since it is not a mutation process.

We could allow many readers to read simultaneously. But for the writing operation, we want to block the readers/writers. The writer should be given preference first in case of a lock by other writers. This would prevent a writer from waiting for readers to complete. This is usually referred to as Read Write Mutual Exclusion.

package main

import (
    "fmt"
    "sync"
    "time"
)

func reader(id int, count *int, wg *sync.WaitGroup, mutex *sync.Mutex) {
    defer wg.Done()
    if !mutex.TryLock() {
        fmt.Printf("Reader %d blocked!\n", id)
        return
    }
    defer mutex.Unlock()
    fmt.Printf("Reader %d: read count %d\n", id, *count)
}

func writer(id, increment int, count *int, wg *sync.WaitGroup, mutex *sync.Mutex) {
    defer wg.Done()
    defer mutex.Unlock()
    mutex.Lock()
    *count += increment
    time.Sleep(5 * time.Millisecond)
    fmt.Printf("Writer %d: wrote count %d\n", id, *count)
}

func main() {

    count := 1
    var wg sync.WaitGroup
    var mutex sync.Mutex
    readers := 5
    writers := 3
    wg.Add(readers)
    for i := 0; i < readers; i++ {
        go reader(i, &count, &wg, &mutex)
    }

    wg.Add(writers)
    for i := 0; i < writers; i++ {
        go writer(i, 1, &count, &wg, &mutex)
    }
    wg.Wait()

}
Enter fullscreen mode Exit fullscreen mode
$ go run --race no-rwmutex.go 
Reader 0: read count 1
Reader 3 blocked!
Reader 2 blocked!
Reader 1 blocked!
Reader 4 blocked!
Writer 0: wrote count 2
Writer 1: wrote count 3
Writer 2: wrote count 4

$ go run --race no-rwmutex.go 
Reader 2 blocked!
Reader 0: read count 1
Reader 1 blocked!
Reader 4 blocked!
Reader 3 blocked!
Writer 2: wrote count 2
Writer 0: wrote count 3
Writer 1: wrote count 4
Enter fullscreen mode Exit fullscreen mode

The above example has a reader and a writer method, the reader method simply has to read the shared resource count. It acquires a Mutex lock before reading and unlocks it thereafter. Similarly, the writer function is used for incrementing the count shared resource.

The reader method has a TryLock method that tries to acquire a mutex lock on the resource, if the resource is already locked, the function will return false and hence we can say that the reading is blocked(just for demonstration). And if the function TryLock returns true, it will acquire the Lock. We further defer the Unlock and access the count variable which is passed as a reference.

The writer method is simply acquiring the Lock and incrementing the counter and thereafter Unlock is called with defer.

In the above code:

  • The reader and the writer both might be waiting for the lock to be released, however, for readers to wait for reading doesn't make sense.
  • Because if you would want to just read a particular memory address, there shouldn't be any locks for one reader to wait for other readers to finish.
  • However, for writing, there has to be a lock. The mutex lock will lock the resource irrespective of the reader or writer.

This might not be as visible here, but it might be the reason, that all the readers are blocked from reading due to another reader's or writer's lock.

The sync package has the RWMutex that does this exact same thing. It is almost similar to the Mutex however, it would allow concurrent reading operation and prefer writing operation before readers to prevent writer starvation.

package main

import (
    "fmt"
    "sync"
    "time"
)

func reader(id int, count *int, wg *sync.WaitGroup, mutex *sync.RWMutex) {
    defer wg.Done()
    defer mutex.RUnlock()

    if !mutex.TryRLock() {
        fmt.Printf("Reader %d blocked!\n", id)
        return
    }
    fmt.Printf("Reader %d: read count %d\n", id, *count)

}

func writer(id, increment int, count *int, wg *sync.WaitGroup, mutex *sync.RWMutex) {
    defer wg.Done()
    defer mutex.Unlock()
    mutex.Lock()

    *count += increment
    time.Sleep(5 * time.Millisecond)
    fmt.Printf("Writer %d: wrote count %d\n", id, *count)
}

func main() {

    count := 1
    var wg sync.WaitGroup
    var mutex sync.RWMutex
    readers := 5
    writers := 3
    wg.Add(readers)
    for i := 0; i < readers; i++ {
        go reader(i, &count, &wg, &mutex)
    }

    wg.Add(writers)
    for i := 0; i < writers; i++ {
        go writer(i, 1, &count, &wg, &mutex)
    }
    wg.Wait()

}
Enter fullscreen mode Exit fullscreen mode
$ go run --race rwmutex.go 
Reader 0: read count 1
Reader 3: read count 1
Reader 1: read count 1
Reader 2: read count 1
Reader 4: read count 1
Writer 1: wrote count 2
Writer 0: wrote count 3
Writer 2: wrote count 4
Writer 3: wrote count 5
Writer 4: wrote count 6

$ go run --race rwmutex.go 
Reader 1 blocked!
Reader 0: read count 1
Reader 2 blocked!
Reader 3 blocked!
Reader 4 blocked!
Writer 4: wrote count 2
Writer 0: wrote count 3
Writer 1: wrote count 4
Writer 3: wrote count 5
Writer 2: wrote count 6

Enter fullscreen mode Exit fullscreen mode

In the modified example, all the logic remains the same, just the sync.Mutex is replaced with sync.RWMutex. Also for trying to acquire the lock in the reader method TryRLock is used instead of TryLock which will check if the existing lock acquired is of a reader or writer, if it is a reader, it will return true, else false. Also the Unlock is replaced with the RUnlock method for releasing the read lock. In the writer method, everything remains the same so the writer has to acquire the lock irrespective of whether the current lock is from the reader/writer, so it is a normal Lock and Unlock.

In the above example, we can see all the read operations sometimes are executed instead of getting blocked. This is due to the RWMutex Lock and Unlock on the read operation/function.

  • When one reader is reading, it can't block other readers.
  • However with simple Mutex, the reader is even blocked when another reader is reading.
  • For the write operation, it will be blocked as usual, so if a writer is performing a write operation and a reader/readers come in, they will be blocked, also if in the meantime, while the resource is locked, another writer comes in, the writer will be given preference instead of waiting for all readers to complete. This prevents writer starvation.

You can see, that the readers are still blocked the majority of the time, however, they are blocked due to the writer locking the resource and not any other reader.

There is a bit more suitable example related to file reading and writing with RWMutex, make sure to check that out to get a more clear understanding of an actual case of using the RWMutex.

Channels

This is a big part and I would like to delve into this topic in a separate post. There are some patterns like fan-in, fan-out, worker-pool, pub-sub, etc. which are really common in web applications and backend systems. These patterns we shall explore in the next article.

Channels are a way to provide a safe and idiomatic way for Goroutines to exchange data and coordinate their execution without resorting to low-level mechanisms like shared memory or explicit locking.

That's it from the 30th part of the series, all the source code for the examples are linked in the GitHub on the 100 days of Golang repository.

100-days-of-Golang GitHub Repository

GitHub logo Mr-Destructive / 100-days-of-golang

Scripts and Resources for the 100 days of golang series

100 Days of Go lang

Go lang is a programming language which is easier to write and even provides type and memory safety, garbage collection and structural typing. It is developed by the Google Team in 2009 and open sourced in 2012. It is a programming language made for Cloud native technologies and web applications. It is a general purpose programming lanugage and hence there are no restrictions on the type of programs you wish to make in it.

Resources to learn Golang

Some famous applications made with GO!

Web apps DevOps tools CLI tools
SoundCloud - Music System Prometheus - Monitoring system and time series database gh-cli - Official Github CLI
Uber - Ride Sharing/Cab booking Webapp
โ€ฆ

References:

Conclusion

From this part of the series, the fundamentals of golang's concurrency model were understood specifically spawning go-routines, synchronously executing go-routines with the help of a wait group, mutex locks, and how to secure concurrent access to a shared resource. In the next part of the series, these concepts will be used in asynchronous communication using channels.

Hopefully, you have got the basics of concurrency in golang cleared from this post. If you have any queries, suggestions, or feedback, please feel free to comment below or contact me on the social handles. Thank you for reading. Happy Coding :)

๐Ÿ’– ๐Ÿ’ช ๐Ÿ™… ๐Ÿšฉ
mr_destructive
Meet Rajesh Gor

Posted on July 31, 2023

Join Our Newsletter. No Spam, Only the good stuff.

Sign up to receive the latest update from our blog.

Related

ยฉ TheLazy.dev

About