The Data Science Lab

Positive and Unlabeled Learning (PUL) Using PyTorch

Dr. James McCaffrey of Microsoft Research provides a code-driven tutorial on PUL problems, which often occur with security or medical data in cases like training a machine learning model to predict if a hospital patient has a disease or not.

A positive and unlabeled learning (PUL) problem occurs when a machine learning set of training data has only a few positive labeled items and many unlabeled items. PUL problems often occur with security or medical data. For example, suppose you want to train a machine learning model to predict if a hospital patient has a disease or not, based on predictor variables such as age, blood pressure, and so on. The training data might have a few dozen instances of items that are positive (class 1 = patient has disease) and many hundreds or thousands of instances of data items that are unlabeled and so could be either class 1 = patient has disease, or class 0 = patient does not have disease.

The goal of PUL is to use the information contained in the dataset to guess the true labels of the unlabeled data items. After the class labels of some of the unlabeled items have been guessed, the resulting labeled dataset can be used to train a binary classification model using any standard machine learning technique, such as k-nearest neighbors classification, neural binary classification, logistic regression classification, naive Bayes classification, and so on.

A good way to see where this article is headed is to take a look at the screenshot of a demo program in Figure 1. The demo uses a 200-item dataset of employee information where the ultimate goal is to classify an employee as an introvert (class 0) or an extrovert (class 1). The dataset is positive and unlabeled: there are just 20 positive (extrovert) employees but the remaining 180 employees are unlabeled and could be either introvert or extrovert.

PUL is challenging and there are several techniques to tackle such problems. The demo program repeatedly (eight times) trains a helper binary classifier using the 20 positive employee data items and 20 randomly selected unlabeled items which are temporarily treated as negative. During the eight model training sessions, information about the unused, unlabeled employee data items is accumulated, in a way that will be explained shortly.

After the eight models have been trained and analyzed, the accumulated information is used to guess the true labels of some of the 180 unlabeled employees. Based on two user-supplied threshold values of 0.30 and 0.90, the PUL system believes it has enough evidence to make intelligent guesses for 57 of the 180 unlabeled employees (32 percent of them).

The true class labels for all 200 employees is known by the demo system. Of the 57 class label guesses, 49 were correct and 8 were incorrect (86 percent accuracy). The demo does not continue by using the now-labeled 97 employee data items (the original 20 positive labeled plus the 57 newly labeled) to create a binary classifier, but that would be the next step in a non-demo scenario.

Figure 1:Positive and Unlabeled Learning (PUL) in Action
[Click on image for larger view.] Figure 1: Positive and Unlabeled Learning (PUL) in Action

This article assumes you have an intermediate or better familiarity with a C-family programming language, preferably Python, and a basic familiarity with the PyTorch code library. The source code for the demo program is a bit too long to present in its entirety in this article, but the complete code and training data are available in the accompanying file download. (The PUL data is embedded in commented-form into the source code).

This article focuses on explaining the key ideas you need to understand in order to analyze and process PUL data to suit your problem scenarios. All normal error checking code has been omitted to keep the main ideas as clear as possible.

To run the demo program, you must have Python and PyTorch installed on your machine. The demo programs were developed on Windows 10 using the Anaconda 2020.02 64-bit distribution (which contains Python 3.7.6) and PyTorch version 1.8.0 for CPU installed via pip. Installation is not trivial. You can find detailed step-by-step installation instructions for this configuration in my blog post.

The PUL Employee Data
The data file of employee information has 200 tab-delimited items. The data looks like:

-2   0.39   0   0   1   0.5120   0   1   0
 1   0.24   1   0   0   0.2950   0   0   1
-2   0.36   1   0   0   0.4450   0   1   0
-2   0.50   0   1   0   0.5650   0   1   0
-2   0.19   0   0   1   0.3270   1   0   0
. . .

The first column is introvert or extrovert, encoded as 1 = positive = extrovert (20 items), and -2 = unlabeled (180 items). The goal of PUL is to intelligently guess 0 = negative, or 1 = positive, for as many of the unlabeled data items as possible.

The other columns in the dataset are employee age (normalized by dividing by 100), city (one of three, one-hot encoded), annual income (normalized by dividing by $100,000), and job-type (one of three, one-hot encoded).

The dataset was artificially constructed so that even numbered items [0], [2], [4], etc. are actually class 0 = negative, and odd numbered items [1], [3], [5], etc. are actually class 1. This allows the PUL system to measure its accuracy. In a non-demo PUL scenario, you usually won't know the true class labels.

The PUL Algorithm
The technique presented in this article is based on a 2013 research paper by F. Mordelet and J.P. Vert, titled "A Bagging SVM to Learn from Positive and Unlabeled Examples". That paper uses a SVM (support vector machine) binary classifier to analyze unlabeled data. This article uses a neural binary classifier instead.

In pseudo-code:

create a 40-item train dataset with all 20 positive
  and 20 randomly selected unlabeled items that
  are temporarily treated as negative
    
loop several times
  train a binary classifier using the 40-item train data
  use trained model to score the 160 unused unlabeled
    data items
  accumulate the p-score for each unused unlabeled item
    
  generate a new train dataset with the 20 positive
    and 20 different unlabeled items treated as negative
end-loop
  

for-each of the 180 unlabeled items
  compute the average p-value 

  if avg p-value < lo threshold
    guess its label as negative
  else-if avg p-value > hi threshold
    guess its label as positive
  else
    insufficient evidence to make a guess
  end-if
end-for

Each time through the training loop, the binary classifier will make fairly poor predictions but the average prediction for all iterations will likely be good. Recall that a neural binary classifier will predict by generating a p-value (pseudo-probability) between 0.0 and 1.0 where a p-value less than 0.5 indicates class 0 = negative, and a p-value greater than 0.5 indicates class 1 = positive. Suppose that an unlabeled data item is not used as part of the training data, three times. And suppose that the trained model scores that unlabeled data item as 0.65, 0.22, 0.58 which mean that the unlabeled item was predicted to be class 1 = positive twice, and class 0 = negative once. The average p-value for the item is (0.65 + 0.32 + 0.78) / 3 = 0.58. Because the average p-value of the unlabeled item is greater than 0.5, it is most likely class 1 = positive.

If you use a decision threshold of 0.5, every unlabeled data item will be guessed as positive or negative. However, many of the guesses where the average p-value is close to 0.5 will likely be incorrect. An alternative approach taken by the demo is to only guess labels where the average p-value is below a low threshold (0.3) or above a high threshold (0.90). Items with average p-values between 0.30 and 0.90 are judged to be ambiguous so no label is guessed.

The low and high threshold values are system hyperparameters that must be determined by the nature of your problem scenario. Adjusting the threshold values towards 0.5 will increase the number of guesses for the unlabeled data items, but probably decrease the accuracy of those guesses.

Generating a Dynamic Dataset
Somewhat unexpectedly, the most difficult part of a PUL system is wrangling the data to generate dynamic (changing) training datasets. The challenge is to be able to create an initial training dataset with the 20 positive items and 20 randomly selected unlabeled items like so:

train_file = ".\\Data\\employee_pul_200.txt"
train_ds = EmployeeDataset(train_file, 20, 180)

And then inside a loop, be able to reinitialize the training dataset with the same 20 positive items but 20 different unlabeled items:

train_ds.reinit()

The dynamic dataset architecture used by the demo program is illustrated in Figure 2. The diagram shows a 14-item dummy PUL dataset with four positive items and ten unlabeled items rather than the 200-item demo dataset. The source PUL data is read into memory as four Python lists of arrays. The first two lists hold the predictors and labels for the four positive items. Note that the positive labels don't need to be explicitly stored because they're all 1, but explicit storage make the code easier to work with. The second two lists hold the predictors and labels for the more numerous unlabeled items, where the unlabeled classes are temporarily all marked as class 0 = negative.

Figure 2: A Dynamic Virtual Dataset for PUL
[Click on image for larger view.] Figure 2: A Dynamic Virtual Dataset for PUL

Each dynamic dataset needs all four of the positive items and four randomly selected unlabeled-marked-as-negative items. It would be inefficient to duplicate data, so all that's needed is information about which rows of the data in memory belong to the positive items and which rows belong to the unlabeled-negative items. And because all positive items are always used in each dynamic dataset, the only information needed is the four rows of unlabeled-negative that are in the dynamic dataset ([6], [3], [8], [2]), and which six rows are not part of the dynamic dataset ([0], [1], [4], [5], [7], [9]).

The virtual dynamic dataset has size 8. To fetch a specified item from it, if the requested index is between [0] and [3] the item can be accessed directly because it must be a positive item. For example, to get the predictor values for virtual item [2], [2] is fetched from memory giving (7.0, 8.0) from Figure 2.

If the requested index is greater than [3] then the requested index must be mapped to its location in memory. For example, to get virtual item [6], the 6 is mapped by subtracting 4 (number of unlabeled-negative items), giving [2]. That value is used to look into the p array that stores memory locations, giving [8] from Figure 2. Item [8] is looked up in memory giving predictor values (11.0, 12.0).

The key takeaway is that PUL systems are not trivial. You must spend a significant amount of engineering time and effort to deal with data wrangling.

The demo code that implements a dynamic virtual dataset for the employee PUL data is presented in Listing 1. As is often the case, data wrangling code is tedious and tricky.

Listing 1: Defining a Dynamic Dataset for PUL Data

class EmployeeDataset(T.utils.data.Dataset):
  # label  age   city   income   job-type
  #   1    0.39  1 0 0  0.5432   1 0 0
  #  -2    0.29  0 0 1  0.4985   0 1 0  (unlabeled)
  # . . .
  #  [0]   [1]  [2 3 4]   [5]   [6 7 8]

  def __init__(self, fn, tot_num_pos, tot_num_unl):
    self.rnd = np.random.RandomState(1)

    self.tot_num_pos = tot_num_pos  # number positives
    self.tot_num_unl = tot_num_unl  # num unlabeleds

    pos_x_lst = []; pos_y_lst = []  # lists of vectors
    unl_x_lst = []; unl_y_lst = []

    ln = 0  # line number (not including comments)
    j = 0  # counter for unlabeleds

    self.unl_idx_to_line_num = dict()
    # key = idx of an unlabeled item in memory,
    # val = corresponding line number in src data file

    fin = open(fn, "r")  # four lists of arrays
    for line in fin:
      line = line.strip()
      if line.startswith("#"): continue

      arr = np.fromstring(line, sep="\t", \
        dtype=np.float32)
      if arr[0] == 1:
        pos_x = arr[[1,2,3,4,5,6,7,8]]
        pos_y = 1  # always 1 
        pos_x_lst.append(pos_x)
        pos_y_lst.append(pos_y)
      elif arr[0] == -2:  # unlabeled
        unl_x = arr[[1,2,3,4,5,6,7,8]]
        unl_y = 0  # treat unlabeleds as negative
        unl_x_lst.append(unl_x)
        unl_y_lst.append(unl_y)
        self.unl_idx_to_line_num[j] = ln
        j +=1
      else:
        print("Fatal: unknown label in file")

      ln += 1  # only data lines

    fin.close()
 
    # data actual storage in 4 tensor-arrays
    self.train_x_pos = T.tensor(pos_x_lst, \
      dtype=T.float32) # predictors for positives
    self.train_y_pos = T.tensor(pos_y_lst, \
      dtype=T.float32).reshape(-1,1) # positives (1s)
    self.train_x_unl = T.tensor(unl_x_lst, \
      dtype=T.float32) # predictors for unlabels
    self.train_y_unl = T.tensor(unl_y_lst, \
      dtype=T.float32).reshape(-1,1) 

    self.num_pos_unl = 2 * tot_num_pos 

    # indices of active and inactive unlabeled items
    all_unl_indices = np.arange(tot_num_unl)  # 180
    self.rnd.shuffle(all_unl_indices)
    self.p = all_unl_indices[0 : tot_num_pos]  # 20 
    self.q = all_unl_indices[tot_num_pos : tot_num_unl]

  def __len__(self):
    return self.num_pos_unl  # virtual ds size 

  def __getitem__(self, idx):
    if idx < self.tot_num_pos:  # small: fetch directly
      return (self.train_x_pos[idx], self.train_y_pos[idx])
    else:   # large index = an unlabeled = map index
      ofset = idx - self.tot_num_pos
      ii = self.p[ofset]  # index of active unlabeled item
      return (self.train_x_unl[ii], self.train_y_unl[ii])

  def reinit(self):  # get (20) different unlabeled items
    all_unl_indices = np.arange(self.tot_num_unl)
    self.rnd.shuffle(all_unl_indices)
    self.p = all_unl_indices[0 : self.tot_num_pos] 
    self.q = all_unl_indices[self.tot_num_pos : \
      self.tot_num_unl]

The code presented in Listing 1 is specific to the employee PUL data. However, by analyzing it carefully, you will be able to adapt the code to meet your own PUL scenarios.

Overall Program Structure
The overall program structure of the demo is presented in Listing 2. The code is moderately long and complex, but PUL problem scenarios are difficult.

Listing 2: Overall Program Structure

import torch as T
# employee_pul.py
# PyTorch 1.8.0-CPU Anaconda3-2020.02  Python 3.7.6
# Windows 10 

import numpy as np
import torch as T
device = T.device("cpu")  # apply to Tensor or Module

# ----------------------------------------------------------

class EmployeeDataset(T.utils.data.Dataset):
  # see Listing 1

# ----------------------------------------------------------

class Net(T.nn.Module):
  # see Listing 3

# ----------------------------------------------------------

def train(net, ds, bs, me, le, lr, verbose):
  # see Listing 4

# ----------------------------------------------------------

def truth_of_line(ln):
  # actual label for 0-based line number of PUL file
  # in non-demo, you'd have to compute
  if ln % 2 == 0: return 0  # files set up this way 
  else: return 1

# ----------------------------------------------------------

def main():
  # 0. get started
  print("\nEmployee PUL using PyTorch \n")
  T.manual_seed(1)
  np.random.seed(1)

  # 1. create data objects
  print("Creating dynamic Employee train Dataset ")
  print("Dataset has 20 positive and 180 unlabeled ")
  train_file = ".\\Data\\employee_pul_200.txt"
  train_ds = EmployeeDataset(train_file, 20, 180)

  # 2. create neural network
  print("\nCreating 8-(10-10)-1 binary NN classifier ")
  net = Net().to(device)

  # 3. prepare for training multiple times
  print("\nSetting training parameters \n")
  bat_size = 10
  lrn_rate = 0.01
  max_epochs = 800
  ep_log_interval = 100

  print("batch size = " + str(bat_size))
  print("lrn_rate = %0.2f " % lrn_rate)
  print("max_epochs = " + str(max_epochs))
  print("loss function = BCELoss() ")
  print("optimizer = SGD \n")

  # track number times each inactive unlabeled is evaluated
  # accumulate sum of p-values from each evaluation
  eval_counts = np.zeros(180, dtype=np.int64)
  eval_sums = np.zeros(180, dtype=np.float32)

# ----------------------------------------------------------

  # 4. accumulate p-values for inactive items after session
  num_trials = 8  # number times to train on a subset
  for trial in range(num_trials):
    print("Training model " + str(trial) + " of " + \
str(num_trials), end="")
    train(net, train_ds, bat_size, max_epochs, \
      ep_log_interval, lrn_rate, verbose=False) 

    print("  Done. Scoring inactive unlabeled items ")
    net.eval()
    for i in train_ds.q:  # idxs of inactive unlabeleds
      x = train_ds.train_x_unl[i]  # predictors
      with T.no_grad():
        p = net(x)          # between 0.0 and 1.0
      eval_counts[i] += 1
      eval_sums[i] += p.item()

    train_ds.reinit()   # get different unlabeleds

# ----------------------------------------------------------

  # 5. guess 0 or 1 labels for unlabeled items
  print("\nGuessing 0 or 1 for unlabeled items ")

  lo = 0.30; hi = 0.90   # tune for accuracy vs. quantity

  # there is more information about positives so
  # to label an unknown as positive you need a higher 
  # p-value threshold criterion.

  print("pseudo-prob thresholds: %0.2f  %0.2f " % (lo, hi))

  num_correct = 0; num_wrong = 0
  for i in range(180):  # process each unlabeled data item
    ln = train_ds.unl_idx_to_line_num[i]  # line num in PUL file

    if eval_counts[i] == 0:
      print("Fatal: Never evaluated this unlabeled item ")
      input()
    else:
      avg_p = (eval_sums[i] * 1.0) / eval_counts[i]

      if avg_p >= lo and avg_p <= hi:  # too close to 0.5
        pass
      elif avg_p < lo and truth_of_line(ln) == 0:  # even class 0
        num_correct += 1
      elif avg_p > hi and truth_of_line(ln) == 1:  # odd class 1
        num_correct += 1
      else:
        num_wrong += 1

  print("\n---------------\n")
  num_guessed = num_correct + num_wrong
  print("num labels guessed = " + str(num_guessed))
  print("num correct guessed labels = " + str(num_correct))
  print("num wrong guessed labels   = " + str(num_wrong))
  acc = (1.0 * num_correct) / (num_correct + num_wrong) 
  pct = (1.0 * (num_correct + num_wrong)) / 180

  print("pct of unlabeled items guessed = %0.4f " % pct)
  print("accuracy of guessed items = %0.4f " % acc)

  print("\nEnd PUL demo ")

# ----------------------------------------------------------

if __name__ == "__main__":
  main()

The demo begins by setting system random number generators so that program runs are reproducible:

def main(): 
  T.manual_seed(1)
  np.random.seed(1)
. . .

Next, an initial version of the employee dataset is created, and the binary classifier is instantiated:

  train_file = ".\\Data\\employee_pul_200.txt"
  train_ds = EmployeeDataset(train_file, 20, 180)
  net = Net().to(device)

Training is prepared using these statements:

  bat_size = 10
  lrn_rate = 0.01
  max_epochs = 800
  ep_log_interval = 100
  eval_counts = np.zeros(180, dtype=np.int64)
  eval_sums = np.zeros(180, dtype=np.float32)
  num_trials = 8  # number times to train on a subset

The eval_counts array holds the number of times each of the 180 unlabeled items is not used as part of the training data and is therefore scored by the current binary classification model. The eval_sums array holds the accumulated p-values for each unlabeled item (when it's not used as part of the training data and is scored).

The main loop is:

  for trial in range(num_trials):
    train(net, train_ds, bat_size, max_epochs, \
      ep_log_interval, lrn_rate, verbose=False) 

    print("  Done. Scoring inactive unlabeled items ")
    net.eval()
    for i in train_ds.q:  # idxs of inactive unlabeleds
      x = train_ds.train_x_unl[i]  # predictors
      with T.no_grad():
        p = net(x)          # between 0.0 and 1.0
      eval_counts[i] += 1
      eval_sums[i] += p.item()

    train_ds.reinit()   # get different unlabeleds

After a binary classifier is trained using a dynamic dataset that has all 20 positive items and 20 randomly selected unlabeled data items that have been marked as negative, the code walks through the unlabeled items that were not part of the training data. Each of these unused unlabeled items is scored.

After training has occurred several times (8 in the demo), the accumulated p-values are analyzed to guess the class of each unlabeled item:

  for i in range(180):  # process each unlabeled data item
    ln = train_ds.unl_idx_to_line_num[i]  # line num

    if eval_counts[i] == 0:
      print("Fatal: Never evaluated this unlabeled item ")
      input()
    else:
      avg_p = (eval_sums[i] * 1.0) / eval_counts[i]

      if avg_p >= lo and avg_p <= hi:  # too close to 0.5
        pass
      elif avg_p < lo and truth_of_line(ln) == 0:  # class 0
        num_correct += 1
      elif avg_p > hi and truth_of_line(ln) == 1:  # class 1
        num_correct += 1
      else:
        num_wrong += 1

If an eval_count value for an unlabeled item is 0, that means the item was never left out of the training data, which means it was part of the randomly selected unlabeled items in the training dataset on every training iteration. This is statistically nearly impossible, and almost certainly indicates a logic error.

The Binary Classifier
The definition of the neural binary classifier used by the PUL system is presented in Listing 4.

Listing 4: Neural Binary Classifier Definition

class Net(T.nn.Module):
  # binary classifier for Employee data

  def __init__(self):
    super(Net, self).__init__()
    self.hid1 = T.nn.Linear(8, 10)  # 8-(10-10)-1
    self.hid2 = T.nn.Linear(10, 10)
    self.oupt = T.nn.Linear(10, 1)

    T.nn.init.xavier_uniform_(self.hid1.weight) 
    T.nn.init.zeros_(self.hid1.bias)
    T.nn.init.xavier_uniform_(self.hid2.weight) 
    T.nn.init.zeros_(self.hid2.bias)
    T.nn.init.xavier_uniform_(self.oupt.weight) 
    T.nn.init.zeros_(self.oupt.bias)

  def forward(self, x):
    z = T.tanh(self.hid1(x))
    z = T.tanh(self.hid2(z))
    z = T.sigmoid(self.oupt(z)) 
    return z

The classifier accepts 8 input values -- the predictors of age, city (3), income, and job-type (3) -- and emits a single value between 0.0 and 1.0 because sigmoid() activation is used on the output node. The classifier has two hidden layers, each with 10 nodes, and with tanh() activation. The number of hidden layers, the number of nodes in each layer, and the activation function for each layer, are all hyperparameters that must be determined by trial and error guided by experience.

The classifier uses explicit weight and bias initialization rather than allowing the PyTorch system to supply default initialization values. Neural weight and bias initialization can often have a big impact (good or bad) on model classification accuracy and performance.

Training the Binary Classifier
The demo program encapsulates the training code into a single train() function. The definition of train() is presented in Listing 5.

Listing 5: Training the Binary Classifier

def train(net, ds, bs, me, le, lr, verbose):
  # NN, dataset, batch_size, max_epochs,
  # log_every, learn_rate. optimizer and loss hard-coded.
  net.train()
  data_ldr = T.utils.data.DataLoader(ds, batch_size=bs,
    shuffle=True)
  loss_func = T.nn.BCELoss()  # assumes sigmoid activation
  opt = T.optim.SGD(net.parameters(), lr=lr)
  for epoch in range(0, me):
    epoch_loss = 0.0
    for (batch_idx, batch) in enumerate(data_ldr):
      X = batch[0]  # inputs
      Y = batch[1]  # targets

      opt.zero_grad()                # prepare gradients
      oupt = net(X)                  # compute output/target
      loss_val = loss_func(oupt, Y)  # a tensor
      epoch_loss += loss_val.item()  # accumulate for display
      loss_val.backward()            # compute gradients
      opt.step()                     # update weights

    if epoch % le == 0 and verbose == True:
      print("epoch = %4d   loss = %0.4f" % (epoch, epoch_loss))

The training code is ordinary in the sense that there is nothing special needed for a PUL scenario. The train() method uses SGD optimization (stochastic gradient descent). This is the most rudimentary optimization technique. For complex data with many features and complex neural binary classification architecture, the Adam optimizer often gives better results.

Wrapping Up
When you have a dataset with just positive (class 1) and unlabeled (could be class 0 or class 1) data items, there is no magic technique. You must intelligently guess the true class labels for the unlabeled items using information in the dataset. Because the process of labeling training data for a machine learning model is often costly and time consuming, PUL problem scenarios are becoming increasingly common.

Because the PUL guessing process is probabilistic, there are many approaches you can use. The approach presented in this article is based on a deep neural binary classifier, and is new and mostly unexplored. Most alternative techniques for guessing class labels for PUL data use some form of k-means clustering. The idea is that unlabeled data items that are close to positive class items are more likely to be positive than negative, and that a cluster of data items that are all unlabeled, and which are all far from a cluster of mostly positive data items, are most likely negative.

The PUL guessing approaches that use k-means all assume that the distance between two data items can be measured (usually by Euclidean distance), but this means that all data items must be strictly numeric. The technique presented in this article has the advantage that it can work with both numeric and categorical data (or mixed numeric and categorical as in the demo data).

comments powered by Disqus

Featured

  • Compare New GitHub Copilot Free Plan for Visual Studio/VS Code to Paid Plans

    The free plan restricts the number of completions, chat requests and access to AI models, being suitable for occasional users and small projects.

  • Diving Deep into .NET MAUI

    Ever since someone figured out that fiddling bits results in source code, developers have sought one codebase for all types of apps on all platforms, with Microsoft's latest attempt to further that effort being .NET MAUI.

  • Copilot AI Boosts Abound in New VS Code v1.96

    Microsoft improved on its new "Copilot Edit" functionality in the latest release of Visual Studio Code, v1.96, its open-source based code editor that has become the most popular in the world according to many surveys.

  • AdaBoost Regression Using C#

    Dr. James McCaffrey from Microsoft Research presents a complete end-to-end demonstration of the AdaBoost.R2 algorithm for regression problems (where the goal is to predict a single numeric value). The implementation follows the original source research paper closely, so you can use it as a guide for customization for specific scenarios.

  • Versioning and Documenting ASP.NET Core Services

    Building an API with ASP.NET Core is only half the job. If your API is going to live more than one release cycle, you're going to need to version it. If you have other people building clients for it, you're going to need to document it.

Subscribe on YouTube