CV Chapter 3 Segmentation

Questions about the lecture 'Computer Vision' of the RWTH Aachen Chapter 3 Segmentation

Questions about the lecture 'Computer Vision' of the RWTH Aachen Chapter 3 Segmentation


Kartei Details

Karten 53
Sprache English
Kategorie Informatik
Stufe Universität
Erstellt / Aktualisiert 04.02.2017 / 19.02.2017
Weblink
https://card2brain.ch/box/20170204_cv_chapter_3_segmentation
Einbinden
<iframe src="https://card2brain.ch/box/20170204_cv_chapter_3_segmentation/embed" width="780" height="150" scrolling="no" frameborder="0"></iframe>

What is the motivation of segmentation? [4]

1. Image regions

2. Video frames into shots

3. Figure-ground

4. Object-level

What are the basic questions of segmentation? [2]

1. What things should be grouped?

2. What cues indicate groups?

What are the characteristics of the Gestalt school? [3]

1. “Grouping key to visual perception”

2. Elements of collection can have properties resulting from relations

3. “Whole is greater than sum of its parts”

What are the Gestalt factors? [8]

1. Not grouped // . . . .

2. Proximity // .. .. .. ..

3. Similarity// ***..***

4. Common fate and region

5. Parallelism // << || >>

6. Symmetry // >< || <>

7. Continuity// Edges

8. Closure // o <>

What are difficulties regarding Gestalt factors? [2]

1. Difficult for algorithms

2. Occlusion covers above factors

What is the basic assumption for image segmentation?

Group of pixel that belong together

What are the characteristics of grouping by pixel intensity? [3]

1. Find representative centers and map pixels to nearest

2. Center minimizing SSD Sum|p-c|²

3. Center → Membership and Membership → Center

What is the definition of K-Means algorithm? [3]

1. Randomly initialize k cluster center

2. Determine points for centers

3. If center is false for points update and repeat

What are the issues of K-Means? [2]

1. Will find local but not always global minimum

2. Choose next proportional to |p-c|² till k centers with error O(log k)*optimal

What are the advantages of K-Means? [2]

1. Simple and fast

2. Converges to local minimum within cluster squared error

What are the disadvantages of K-Means? [4]

1. Which k? // NP-hard for even k=2

2. Sensitive to initial centers and outliers

3. Detects only convex clusters

4. Assuming means

What are possible feature spaces? [3]

1. 1D intensity similarity // Extended by position

2. 3D color similarity

3. 24D texture similarity // Possibilities from black to white

What are the basic questions of probabilistic clustering? [2]

1. What is the probability that point p belongs to cluster c?

2. What is the shape of the cluster c?

What is the basic assumption for generative model?

Assume that pixel are generated by continuous function

What are the characteristics of mixture of gaussians (MoG)? [4]

1. K Gaussian blobs with means mye_j cov matrices Sigma_j and dim D

2. p(x|theta_j) = 1/2pi^(D/2)*|Sigma_j|^(1/2) * exp(-1/2 *(x-mye_)^T *Sigma_j^-1 *(x-mye_j) )

3. Select blob j with probability pi_j

4. Likelihood of x is weighted mixture of Gaussians p(x|theta) = Sigma_j^K pi_j(x|theta_j)

What are the characteristics of expectation maximization (EM)? [3]

1. Maximize p(data|theta) = Prod_n^N p(x_n|theta)

 

2. Current blobs → Ownership // Expectation

3. Ownership probabilities → Update blobs // Maximization

What are the characteristics of expectation?

gamma_j(xn_) ← value_j / Sigma_k^K value_k with value_j =pi_j * N(x_n|mye_j,Sigma_j)

What are the characteristics of maximization? [4]

1. N_j ← Sum_n^N gamma_j(x_n)

2. pi_j^new ← N_j/N

3. mye_j^new ← 1/N_j * Sigma_n^N gamma_j(x_n)*x_n

4. Sigma_j^new ← 1/N_j * Sigma_n^N gamma_j(x_n)*(x_n-mye_j^new)^(T+1)

What is expectation maximization useful for? [3]

1. Any clustering problem // Any model estimation problem

2. Missing data problem and finding outliers

3. segmentation problems color motion fore/background

What are the advantages of expectation maximization? [3]

1. Soft assignments between data and cluster

2. Generative model can predict novel points

3. Compact storage

What are the disadvantages of expectation maximization? [3]

1. Initialization // With some k-means iterations

2. Unknown number of components // Solution with model selection (AIC, BIC), Dirichlet process mixture

3. Which generative model?

What is the basic idea of mean-shift segmentation? [2]

1. Modes in histogram are local maximum of the density

2. Cluster is set of points in attraction of one mode

What is the mean-shift segmentation algorithm? [3]

1. Initialize random seed and window W

2. Mean of W Sigma_xinW x*H(x)

3. Shift W to mean and redo 2. till convergence

What are the characteristics of real modality analysis? [3]

1. Choose multiple windows

2. Run mean-shift segmentations parallel

3. Label the traversed points

Which speedups for mean-shift segmentation? [2]

1. Label all points choose in radius r

2. Label all points in search path

What are the advantages of mean-shift segmentation? [4]

1. Model-free, no prior shape assumption

2. Only one parameter window size

3. Finds number of modes

4. Robust to outliers

What are the disadvantages of mean-shift segmentation? [2]

1. Output depends on not trivial window size

2. Expensive (2s/image)

What is the idea of markow random fields (MRF)? [3]

1. Rich probabilistic models for images

2. Lear local effects and get global effects out

3. Goal is to infer optimal labeling of MRF

What are the components of markov random fields (MRF)? [4]

1. Observed evidence on y-plane the image

2. Image scene compatibility function Phi(xi,yi)

3. Hidden true states on x-plane the scene

4. Neighborhood relations or scene scene compatibility function Psi(xi,xj)

What is the definition of the network joint probability p(x,y) of MRF?

Network joint probability p(x,y)=Prod_i Phi(xi,yi)* Prod_ij Psi(xi,xj)

What is the alternative of maximizing the network joint probability?

Minimizing -log p(x,y)=-Sum_i log(Phi(xi,yi)) - Sum_ij log(Psi(xi,xj))

What is the definition of the energy E(x,y) of MRF?

E(x,y) = Sum_i phi(xi,yi)) + Sum_ij psi(xi,xj)

What are the potentials of the energy of MRF? [2]

1. Unary or single-node phi encoding local info // membership of pixel to a class

2. Pairwise psi encoding neighborhood info

Which possibilities exist to set the potentials of the energy of MRF? [2]

1. Color modeled with MoG

2. Edge with contrast sensitive Potts model

Which inference algorithms for MRF exist? [6]

1. Gibbs sampling

2. Simulated annealing

3. Iterated conditional modes (ICM)

4. Variational methods

5. Belief propagation

6. Graph cuts

What are the characteristics of graph cuts for MRF? [4]

1. Popular fast tool for certain class of energy functions

2. Convert MRF into source-sink graph (SSG)

3. Minimum cost cut with max-flow/min-cut in polynomial time

4. Only minimizes submodular binary energies (Boros and Hummer 2002)

What is the definition of submodular of graph cut for MRF? [3]

1. E(s,s)+E(t,t)\leqE(s,t)+E(t,s)

2. Discrete equivalent to convexity

3. Every local is global minimum

What are the formulas of graph cut for MRF? [2]

1. wij = exp(- Diij/2*sigma²)

2. E(x,y) = Sum_i phi_i(xi) + Sum_ij wij * delta(xi!=xj)

What holds for the graph cut algorithm for MRF? [4]

1. Expect phi_i(s or t) ~ exp(-|Ii-(Is or It)|² /2*sigma²)

2. Is It are given expected intensities and can be re-estimated

3. Hard constrains are not required in general

4. Regional bias based on any intensity model with phi_i(Li)=-log p(Ii|Li) and Li s or t

What is the definition of the s-t-Mincut problem? [4]

1. Graph with vertices edges and costs

2. st-cut divides nodes between source and sink

3. Cost is sum of edges going from source to sink region

4. st-mincut has minimal cost