We love eBooks
    Download Support Vector Machines and Perceptrons: Learning, Optimization, Classification, and Application to Social Networks (SpringerBriefs in Computer Science) pdf, epub, ebook
    Publisher

    This site is safe

    You are at a security, SSL-enabled, site. All our eBooks sources are constantly verified.

    Support Vector Machines and Perceptrons: Learning, Optimization, Classification, and Application to Social Networks (SpringerBriefs in Computer Science)

    By M.N. Murty

    What do you think about this eBook?

    About

    This work reviews the state of the art in SVM and perceptron classifiers. A Support Vector Machine (SVM) is easily the most popular tool for dealing with a variety of machine-learning tasks, including classification. SVMs are associated with maximizing the margin between two classes. The concerned optimization problem is a convex optimization guaranteeing a globally optimal solution. The weight vector associated with SVM is obtained by a linear combination of some of the boundary and noisy vectors. Further, when the data are not linearly separable, tuning the coefficient of the regularization term becomes crucial. Even though SVMs have popularized the kernel trick, in most of the practical applications that are high-dimensional, linear SVMs are popularly used. The text examines applications to social and information networks. The work also discusses another popular linear classifier, the perceptron, and compares its performance with that of the SVM in different application areas.>

    Download eBook Link updated in 2017
    Maybe you will be redirected to source's website
    Thank you and welcome to our newsletter list! Ops, you're already in our list.

    Related to this eBook

    Browse collections Find similar eBooks

    Keep connected to us

    Follow us on Social Media or subscribe to our newsletter to keep updated about eBooks world.

    Explore eBooks

    Browse all eBook collections

    Collections is the easy way to explore our eBook directory.